Unnamed: 0
int64
0
7.24k
id
int64
1
7.28k
raw_text
stringlengths
9
124k
vw_text
stringlengths
12
15k
1,100
2,003
Incremental A  S. Koenig and M. Likhachev Georgia Institute of Technology College of Computing Atlanta, GA 30312-0280 skoenig, mlikhach @cc.gatech.edu  Abstract Incremental search techniques find optimal solutions to series of similar search tasks much faster than is possible by solving each search task from scratch. While researchers have developed incremental versions of uninformed search methods, we develop an incremental version of A*. The first search of Lifelong Planning A* is the same as that of A* but all subsequent searches are much faster because it reuses those parts of the previous search tree that are identical to the new search tree. We then present experimental results that demonstrate the advantages of Lifelong Planning A* for simple route planning tasks. 1 Overview Artificial intelligence has investigated knowledge-based search techniques that allow one to solve search tasks in large domains. Most of the research on these methods has studied how to solve one-shot search problems. However, search is often a repetitive process, where one needs to solve a series of similar search tasks, for example, because the actual situation turns out to be slightly different from the one initially assumed or because the situation changes over time. An example for route planning tasks are changing traffic conditions. Thus, one needs to replan for the new situation, for example if one always wants to display the least time-consuming route from the airport to the conference center on a web page. In these situations, most search methods replan from scratch, that is, solve the search problems independently. Incremental search techniques share with case-based planning, plan adaptation, repair-based planning, and learning search-control knowledge the property that they find solutions to series of similar search tasks much faster than is possible by solving each search task from scratch. Incremental search techniques, however, differ from the other techniques in that the quality of their solutions is guaranteed to be as good as the quality of the solutions obtained by replanning from scratch. Although incremental search methods are not widely known in artificial intelligence and control, different researchers have developed incremental search versions of uninformed search methods in the algorithms literature. An overview can be found in [FMSN00]. We, on the other hand, develop an incremental version of A*, thus combining ideas from the algorithms literature and the artificial intelligence literature. We call the algorithm Lifelong Planning A* (LPA*), in analogy to ?lifelong learning? [Thr98], because it reuses  We thank Anthony Stentz for his support. The Intelligent Decision-Making Group is partly supported by NSF awards under contracts IIS9984827, IIS-0098807, and ITR/AP-0113881. The views and conclusions contained in this document are those of the authors and should not be interpreted as representing the official policies, either expressed or implied, of the sponsoring organizations and agencies or the U.S. government. information from previous searches. LPA* uses heuristics to focus the search and always finds a shortest path for the current edge costs. The first search of LPA* is the same as that of A* but all subsequent searches are much faster. LPA* produces at least the search tree that A* builds. However, it achieves a substantial speedup over A* because it reuses those parts of the previous search tree that are identical to the new search tree. 2 The Route Planning Task Lifelong Planning A* (LPA*) solves the following search task: It applies to finite graph search problems on known graphs whose edge costs can increase or decrease over time. denotes the finite set of vertices of the graph.    denotes the set of successors of  vertex  . Similarly,      denotes the set of predecessors of vertex  .  !"# %$'& denotes the cost of moving from vertex  to vertex "()*+,  . LPA* always determines a shortest path from a given start vertex ,-/.10324.56 to a given goal vertex 879 0;:  , knowing both the topology of the graph and the current edge costs. We use <>=,  to denote the start distance of vertex ?@ , that is, the length of a shortest path from  -/.102A. to  . To motivate and test LPA*, we use a special case of these search tasks that is easy to visualize. We apply LPA* to navigation problems in known eight-connected gridworlds with cells whose traversability can change over time. They are either traversable (with cost one) or untraversable. LPA* always determines a shortest path between two given cells of the gridworld, knowing both the topology of the gridworld and which cells are currently blocked. This is a special case of the graph search problems on eight-connected grids whose edge costs are either one or infinity. As an approximation of the distance between two cells, we use the maximum of the absolute differences of their x and y coordinates. This results in consistent heuristics that are for eight-connected grids what Manhattan distances are for four-connected grids. 3 Reusing Information from Previous Searches The graph search problems can be solved with traditional graph-search methods, such as breadth-first search, if they update the shortest path every time some edge costs change. They typically do not take advantage of information from previous searches. The following example, however, shows that this can be advantageous. 5CED F Consider the gridworlds of size B shown in Figure 1. The original gridworld is shown on top and the changed gridworld is shown at the bottom. The traversability of only a few cells has changed. In particular, three blocked cells became traversable (namely, B3, C5, and D2) and three traversable cells became blocked (namely, A1, A4, D3). Thus, two percent of the cells changed their status but the obstacle density remained the same. The figure shows the shortest paths in both cases, breaking ties towards the north. Note that we assume that one can squeeze through diagonal obstacles. (This is just an artifact of how we generated the underlying graphs from the mazes.) The shortest path changed since one cell on the original shortest path became blocked. Once the start distances of all cells are known, one can easily trace back a shortest path from the start cell to the goal cell by always greedily decreasing the start distance, starting at the goal cell. This is similar to how A* traces the shortest path back from 79 03: to  -/.10324. using the search tree it has constructed. Thus, we only need to determine the start distances. The start distances are shown in each traversable cell of the original and changed gridworlds. Those cells whose start distances in the changed gridworld have changed from the corresponding ones in the original gridworld are shaded gray. There are two different ways of decreasing the search effort for determining the start distances for the changed gridworld. First, some start distances have not changed and thus need not get recomputed. This is what DynamicSWSF-FP [RR96] does. (DynamicSWSF- Original Eight-Connected Gridworld 1 7 6 4 A 3 3 4 B C D 6 7 8 7 7 2 7 6 3 7 6 5 4 3 2 5 5 4 3 2 2 2 3 4 5 6 4 3 2 1 1 1 sstart 1 9 3 3 2 3 6 6 7 8 5 5 4 5 6 7 8 6 7 7 11 7 6 7 8 8 4 4 7 8 9 10 9 8 7 6 5 5 10 9 8 9 8 7 4 11 10 9 8 5 11 10 11 10 6 5 6 8 8 8 9 8 9 9 11 7 8 7 7 7 8 9 8 8 8 9 10 9 9 9 9 10 10 10 10 10 10 10 11 12 12 13 12 11 11 11 11 11 11 16 15 14 13 16 14 14 14 15 15 15 15 12 sgoal 14 12 13 14 12 12 14 12 13 14 13 14 14 14 15 15 15 15 15 15 15 15 12 16 16 16 16 16 16 16 16 16 Changed Eight-Connected Gridworld 7 6 7 5 5 5 4 6 7 8 7 6 7 6 5 4 3 2 5 3 3 4 7 2 2 3 4 5 6 4 3 2 1 1 1 sstart 1 9 3 3 2 3 6 6 7 8 6 7 7 8 5 5 6 4 5 6 7 8 11 7 6 7 8 4 4 7 8 9 10 9 8 7 6 5 5 10 9 8 9 8 7 11 10 9 8 11 10 14 13 6 12 7 5 6 8 8 8 9 8 9 9 7 7 7 8 9 10 11 8 8 8 9 10 9 9 9 10 11 10 10 10 10 10 11 11 19 18 17 16 12 16 15 14 12 11 11 11 11 19 17 17 17 18 18 18 18 12 sgoal 14 12 13 14 12 12 14 12 13 14 13 14 14 14 15 15 15 15 15 15 15 15 15 19 19 18 17 16 16 16 16 16 Figure 1: Simple Gridworld FP, as originally stated, searches from the goal vertex to the start vertex and thus maintains estimates of the goal distances rather than the start distances. It is a simple matter of restating it to search from the start vertex to the goal vertex. Furthermore, DynamicSWSFFP, as originally stated, recomputes all goal distances that have changed. To avoid biasing our experimental results in favor of LPA*, we changed the termination condition of DynamicSWSF-FP so that it stops immediately after it is sure that it has found a shortest path.) Second, heuristic knowledge, in form of approximations of the goal distances, can be used to focus the search and determine that some start distances need not get computed at all. This is what A* [Pea85] does. We demonstrate that the two ways of decreasing the search effort are orthogonal by developing LPA* that combines both of them and thus is able to replan faster than either DynamicSWSF-FP or A*. Figure 2 shows in gray those cells whose start distances each of the four algorithms recomputes. (To be precise: it shows in gray the cells that each of the four algorithms expands.) During the search in the original gridworld, DynamicSWSF-FP computes the same start distances as breadth-first search during the first search and LPA* computes the same start distances as A*. During the search in the changed gridworld, however, both incremental search (DynamicSWSF-FP) and heuristic search (A*) individually decrease the number of start distances that need to get recomputed compared to breadth-first search, and together (LPA*) decrease the number even more. 4 Lifelong Planning A* Lifelong Planning A* (LPA*) is an incremental version of A* that uses heuristics   to control its search. As for A*, the heuristics approximate  the goal distances of the vertices  . They need to be consistent, that is, satisfy   79 03:  and  $E,; "    " for all vertices  5 and  " 5+   with    79 0;: . LPA* maintains an estimate <   of the start distance <+=  of each vertex  . These values directly correspond to the g-values of an A* search. They are carried forward from search to search. LPA* also maintains a second kind of estimate of the start distances. The rhs-values are one-step lookahead values based on the g-values and thus potentially better informed complete search Original Eight-Connected Gridworld uninformed search heuristic search breadth-first search A* sstart sgoal sstart incremental search DynamicSWSF-FP (with early termination) sstart sgoal sgoal Lifelong Planning A* sstart sgoal complete search Changed Eight-Connected Gridworld sstart uninformed search heuristic search breadth-first search A* sgoal sstart incremental search DynamicSWSF-FP (with early termination) sstart sgoal sgoal Lifelong Planning A* sstart sgoal Figure 2: Performance of Search Methods in the Simple Gridworld than the g-values. They always satisfy the following relationship:     "!$#%'&)( *,+ .-%/021435/6 , if   otherwise. (1) A vertex is called locally consistent iff its g-value equals its rhs-value. This is similar to satisfying the Bellman equation for undiscounted deterministic sequential decision problems. Thus, this concept is important because the g-values of all vertices equal their start distances iff all vertices are locally consistent. However, LPA* does not make every vertex locally consistent. Instead, it uses the heuristics   to focus the search and update only the g-values that are relevant for computing a shortest path from ,-/.10324. to  739 0;: . LPA* maintains a priority queue 7 that always contains exactly the locally inconsistent vertices. These are the vertices whose g-values LPA* potentially needs to update to make them locally consistent. The keys of the vertices in the priority queue correspond to the f-values used by A*, and LPA* always expands the vertex in the priority queue with the smallest key, similar to A* that always expands the vertex in the priority queue with the smallest f-value. By expanding a vertex, we mean executing 10-16 (numbers in brackets refer to line numbers in Figure 3). The key 8  of vertex  is a vector with two components:        * ,+    * * *+ ,+  *+.+*,+    * * ,+ " *,++     !   *,+   * ,+     *"!$#&%'(# + ) * !$#&%'*#  * !#%+'*# + ) +  *-5, +   *,/ .  !$#&%'(# +   *, +    ! 1032 '"4$56-798 *  * +;: * ,5+.+ *,!  + *-,5+ *  *-5, +< .   *-5, +.+ *,= *-5, +.+ *+ *"A"B % C +   *"A"B %C +< .  **A"B % C ++ , **  *-5, +ED * + *  +?@ > *-5, +.+  *-5, +  !  F ,3*-:$5, : + *-,5+ *,+  *-5, +  !  F,3:$: *-,5+3G , *,+  *+ *+  The pseudocode uses the following functions to manage the priority *+ * + queue: U.TopKey returns the smallest priority of all vertices in priority queue . (If is empty, .) U.Pop deletes the vertex with the smallest priority in priority queue and returns the * then + U.TopKey  returns *,+  vertex. U.Insert inserts vertex into priority queue with priority . Finally, U.Remove removes vertex from priority queue .  procedure CalculateKey 01 return      procedure Initialize 02 ; 03 for all 04 05 U.Insert ; ; ; ; procedure UpdateVertex 06 if 07 if U.Remove ; 08 if U.Insert       ; CalculateKey ; procedure ComputeShortestPath 09 while U.TopKey CalculateKey OR 10 U.Pop ; 11 if 12 ; 13 for all UpdateVertex ; 14 else 15 ; 16 for all UpdateVertex      ; procedure Main *+ 17 Initialize ; 18 forever *+ 19 ComputeShortestPath ; 20 Wait for changes in edge * costs; + 21 for all directed edges * with + changed edge costs 22 Update the edge cost ; * + 23 UpdateVertex ; H -,F : H -,F H Figure 3: Lifelong Planning A*. I KJ IEL   M I=N  $O"6   and 8?W   X  RUT&V  <   A    4 where 8QP  SRUT&V  <   A   4    (2)  1 . Keys are compared according to a lexicographic ordering. For example, a key 8(  is smaller than  or equal to a key 8 "/  , denoted by 8(  Z$ Y 8 "  , iff either 8 P  8 P"   or ( 8 P   8 P"   used and 8[W   $ 8 W "   ). 8QP,  corresponds directly to the f-values \  <+=  by A* because both the g-values and rhs-values of LPA* correspond to the g-values of A* and the h-values of LPA* correspond to the h-values of A*. 8 W  corresponds to the g-values of A*. LPA* expands vertices in the order of increasing k P -values and vertices with equal k P -values in order of increasing k W -values. This is similar to A* that expands vertices in the order of increasing f-values (since the heuristics are consistent) and vertices with equal f-values that are on the same branch of the search tree in order of increasing g-values (since it grows the search tree).  A locally inconsistent vertex  is called overconsistent iff <   ^]     . When LPA* expands a locally overconsistent vertex 12-13 , then  +    <+=,  because vertex  has the smallest key among all locally inconsistent vertices.     <+=,  implies that 8  `_ \  baA< =  *c and thus LPA* expands overconsistent vertices in the same order as A*. During the expansion of vertex  , LPA* sets the g-value of vertex  to its rhsvalue and thus its start distance 12 , which is the desired value and also makes the vertex locally consistent. Its g-value then no longer changes until LPA* terminates. A locally     . When LPA* expands inconsistent vertex  is called underconsistent iff < a locally underconsistent vertex 15-16 , then it simply sets the g-value of the vertex to infinity 15 . This makes the vertex either locally consistent or locally overconsistent. If the expanded vertex was locally overconsistent, then the change of its g-value can affect the local consistency of its successors 13 . Similarly, if the expanded vertex was locally underconsistent, then it and its successors can be affected 16 . LPA* therefore updates rhs-values of these vertices, checks their local consistency, and adds them to or removes them from the priority queue accordingly.             LPA* expands vertices until  739 0;: is locally consistent and the key of the vertex to expand next is no smaller than the key of  79 0;: . This is similar to A* that expands vertices until it expands 879 03: at which point in time the g-value of  739 0;: equals its start distance and the f-value of the vertex to expand next is no smaller than the f-value of 79 0;: . It turns out that LPA* expands a vertex at most twice, namely at most once when it is underconsistent and at most once when it is overconsistent. Thus, ComputeShortestPath  returns after a number of vertex expansions that is at most twice the number of vertices. If <  879 0;:  & after the search, then there is no finite-cost path from  - .1032A. to 879 0;: . Otherwise, one can trace back a shortest path from -/.10324. to  79 0;: by always moving from the current vertex  , starting at  79 03: , to any predecessor  " that minimizes <   "#  " ; until  -/.10324. is reached (ties can be broken arbitrarily), similar to what A* can do if it does not use backpointers. The resulting version of LPA* is shown in Figure 3. The main function Main() first calls Initialize() to initialize the search problem 17 . Initialize() sets the initial g-values of all vertices to infinity and sets their rhs-values according to Equation 1 03-04 . Thus, initially  -/.10324. is the only locally inconsistent vertex and is inserted into the otherwise empty priority queue with a key calculated according to Equation 2 05 . This initialization guarantees that the first call to ComputeShortestPath() performs exactly an A* search, that is, expands exactly the same vertices as A* in exactly the same order, provided that A* breaks ties between vertices with the same f-values suitably. Notice that, in an actual implementation, Initialize() only needs to initialize a vertex when it encounters it during the search and thus does not need to initialize all vertices up front. This is important because the number of vertices can be large and only a few of them might be reached during the search. LPA* then waits for changes in edge costs 20 . If some edge costs have changed, it calls UpdateVertex() 23 to update the rhs-values and keys of the vertices potentially affected by the changed edge costs as well as their membership in the priority queue if they become locally consistent or inconsistent, and finally recalculates a shortest path 19 .             5 Optimizations of Lifelong Planning A* There are several simple ways of optimizing LPA* without changing its overall operation. The resulting version of LPA* is shown in Figure 4. First, a vertex sometimes gets removed from the priority queue and then immediately reinserted with a different key. For example, a vertex can get removed on line 07 and then be reentered on line 08 . In this case, it is often more efficient to leave the vertex in the priority queue, update its key, and only change its position in the priority queue. Second, when UpdateVertex  on line 13 computes the rhs-value for a successor of an overconsistent vertex it is unnecessary to take the minimum over all of its respective predecessors. It is sufficient to compute the rhs-value as the minimum of its old rhs-value and the sum of the new g-value of the overconsistent vertex and the cost of moving from the overconsistent vertex to the successor. The reason is that only the g-value of the overconsistent vertex has changed. Since it decreased, it can only decrease the rhs-values of the successor. Third, when UpdateVertex  on line 16 computes the rhs-value for a successor of an underconsistent vertex, the only g-value that has changed is the g-value of the underconsistent vertex. Since it increased, the rhs-value of the successor can only get affected if its old rhs-value was based on the old g-value of the underconsistent vertex. This can be used to decide whether the successor needs to get updated and its rhs-value needs to get recomputed 21? . Fourth, the second and third optimization concerned the computations of the rhs-values of the successors after the g-value of a vertex has changed. Similar optimizations can be made for the computation of the rhs-value of a vertex after the cost of one of its incoming edges has changed.           6 Analytical and Experimental Results We can prove the correctness of ComputeShortestPath().   *+      1  The pseudocode uses the following functions to manage the priority queue: U.Top returns a vertex with the smallest priority *+ * + of all vertices in . (If is empty, then U.TopKey returns .) priority queue * + . U.TopKey  returns the smallest priority of all vertices in priority *  queue +  U.Insert inserts vertex into priority  queue with priority . U.Update *,+ changes the priority of vertex in priority queue to . (It  does nothing if the current priority of vertex already equals .) Finally, U.Remove removes vertex from priority queue .  $ * ,+     * +   *+.+ *,+   * * ,+  *,+.+       !   *+   *,+     *  !$#%+'*# +  ) * !$#%+'*#  * !#%+'*# + ) +     *, *+< *- . ,5+  .   *-5, *-5, +  + *,+ , ! ,  ! + + *-,F *-,= *,+.+ *-,5+.+   *  *-5, +     *, + , ! + *-,+ *+ * * + @> *  A"B % C +   * A"B % C +  .  * A"B % C +.+ , *  *-5, + D * +   *-5, +.+  *-5, +     *, + *-,5+  !Z=,9:(: *-5, + *  .  !$#%+'*# +   *,+    *  *,+ 1 *, +/: *-,F" ,+.+ *,+  B *-,5C5 +   *-, +   =,9:(: *-5, +G ,  !Z *  *,+  B C5 :  *-,F" ,+    , + * * + :* ,+.+ *   .  !#%+'*# +  *,+  ! 0 3 2 '*4(56 !$8     *,+  procedure CalculateKey * *,+  01? return      procedure Initialize 02? ; 03? for all 04? 05? U.Insert procedure UpdateVertex 06? if ( 07? else if 08? else if          ; ; ; ; AND AND AND procedure ComputeShortestPath 09? while U.TopKey CalculateKey 10? U.Top ; 11? if 12? ; 13? U.Remove ; 14? for all 15? if 16? UpdateVertex ; 17? else 18? ; 19? ; 20? for all 21? if 22? if 23? UpdateVertex ;    U.Update CalculateKey ; U.Insert CalculateKey U.Remove ; ; OR ; OR ; *+       procedure Main*+ 24? Initialize ; 25? forever *+ 26? ComputeShortestPath ; 27? Wait for changes in edge * costs; + 28? for all directed with changed edge costs * edges+ 29? ; * + 30? Update the edge ; * cost + 31? if ( *  + )   * + * * + 32? if *  * + * 5+ + 33? else if*  +   * +  34? if * + 35? UpdateVertex ;  : B C5  : -,F H -,F : H -,= 1H :*B -C5H D. : !$-,=#&% 1'*H #  -H  -H   -, ;:"B  C5    -H   *-,5+ :*-,= 1H+.+   -H  . !$#&%'*#  -H    ! 1032 '"4$5 6 8 * * + ;:* 1H+.+ -H ; ; Figure 4: Lifelong Planning A* (optimized version)   Theorem 1 ComputeShortestPath() terminates and one can then trace back a shortest path from ' to  ' by always moving from the current vertex  , starting at  ' , to any predecessor  / that minimizes -% / 21435 / 6  until   is reached (ties can be broken arbitrarily). (The proofs can be found in [LK01].) We now compare breadth-first search, A*, DynamicSWSF-FP, and the optimized version of LPA* experimentally. (We use DynamicSWSF-FP with the same optimizations that we developed for LPA*, to avoid biasing our experimental results in favor of LPA*.) The priority queues of all four algorithms were implemented as binary heaps. Since all algorithms determine the same paths (if they break ties suitably), we need to compare their total search time until a shortest path has been found. Since the actual runtimes are implementation-dependent, we instead use three measures that all correspond to common operations performed by the algorithms and thus heavily influence their runtimes: the total number of vertex expansions   (that is, updates of the g-values, similar to backup operations of dynamic programming for sequential decision problems), the total number of vertex accesses  (for example, to read or change their values), and the total number of heap percolates (exchanges of a parent and child in the heap). Note that we count two vertex expansions, not just one vertex expansion, if LPA* expands the same vertex twice, to avoid biasing our experimental results in favor of LPA*. All of our experiments use fifty eight-connected gridworlds that have size C  and an obstacle density of 40 percent. The start cell is at coordinates (34, 20) and the goal cell is at coordinates (5, 20), where the upper leftmost cell is at coordinates (0, 0). For each gridworld, the initial obstacle configuration was generated randomly. Then, it was changed 500 times in a row, each time by making eight randomly chosen blocked cells traversable and eight randomly chosen traversable cells blocked. Thus, each time one percent of the cells changed their status but the obstacle density remained the same. After each of the 500 changes, the algorithms recomputed a shortest path from the start cell to the goal cell. For each of the four algorithms and each of the three performance measures, the following table reports the mean of the performance measure for the 500 changes: both its average over the fifty mazes and its 95-percent confidence interval over the fifty mazes (assuming a normal distribution with unknown variance). The table confirms the observations made in Section 3: LPA* outperforms the other three search methods according to all three performance measures. complete search ve = va = hp = incremental search ve = va = hp = uninformed search breadth-first search 1331.7 4.4 26207.2 84.0 5985.3 19.7 DynamicSWSF-FP 173.0 4.9 5697.4 167.0 956.2 26.6 heuristic search A* ve = 284.0 5.9 va = 6177.3 129.3 hp = 1697.3 39.9 Lifelong Planning A* ve = 25.6 2.0 va = 1235.9 75.0 hp = 240.1 16.9 We have also applied LPA* successfully to more complex planning tasks, including the kind of route planning tasks that Focussed Dynamic A* [Ste95] applies to. The results will be reported separately. References [FMSN00] D. Frigioni, A. Marchetti-Spaccamela, and U. Nanni. Fully dynamic algorithms for maintaining shortest paths trees. Journal of Algorithms, 34(2):251? 281, 2000. [LK01] M. Likhachev and S. Koenig. Lifelong Planning A* and Dynamic A* Lite: The proofs. Technical report, College of Computing, Georgia Institute of Technology, Atlanta (Georgia), 2001. [Pea85] J. Pearl. Heuristics: Intelligent Search Strategies for Computer Problem Solving. Addison-Wesley, 1985. [RR96] G. Ramalingam and T. Reps. An incremental algorithm for a generalization of the shortest-path problem. Journal of Algorithms, 21:267?305, 1996. [Ste95] A. Stentz. The focussed D* algorithm for real-time replanning. In Proceedings of the International Joint Conference on Artificial Intelligence, pages 1652? 1659, 1995. [Thr98] Sebastian Thrun. Lifelong learning algorithms. In S. Thrun and L. Pratt, editors, Learning To Learn. Kluwer Academic Publishers, 1998.
2003 |@word version:9 advantageous:1 suitably:2 termination:3 d2:1 confirms:1 shot:1 initial:2 configuration:1 series:3 contains:1 document:1 outperforms:1 current:5 subsequent:2 remove:8 update:11 intelligence:4 accordingly:1 constructed:1 predecessor:4 become:1 prove:1 combine:1 planning:20 bellman:1 decreasing:3 actual:3 increasing:4 provided:1 underlying:1 what:4 kind:2 interpreted:1 minimizes:2 developed:3 informed:1 guarantee:1 every:2 expands:14 tie:5 exactly:4 control:3 reuses:3 local:2 topkey:6 path:21 ap:1 might:1 twice:3 initialization:1 studied:1 shaded:1 directed:2 procedure:10 confidence:1 wait:3 get:8 ga:1 influence:1 deterministic:1 center:1 starting:3 independently:1 immediately:2 his:1 coordinate:4 updated:1 heavily:1 programming:1 us:5 satisfying:1 bottom:1 inserted:1 solved:1 connected:9 ordering:1 decrease:4 removed:2 traversable:6 substantial:1 agency:1 broken:2 dynamic:4 motivate:1 solving:3 easily:1 joint:1 recomputes:2 artificial:4 whose:6 heuristic:12 widely:1 solve:4 otherwise:3 favor:3 traversability:2 advantage:2 analytical:1 adaptation:1 relevant:1 combining:1 iff:5 lookahead:1 squeeze:1 parent:1 empty:3 undiscounted:1 produce:1 incremental:15 executing:1 leave:1 develop:2 uninformed:5 solves:1 implemented:1 implies:1 differ:1 successor:10 exchange:1 government:1 generalization:1 insert:8 normal:1 visualize:1 achieves:1 early:2 smallest:7 heap:3 currently:1 replanning:2 individually:1 correctness:1 successfully:1 lexicographic:1 always:11 rather:1 avoid:3 gatech:1 focus:3 check:1 greedily:1 dependent:1 membership:1 typically:1 initially:2 expand:2 overall:1 among:1 denoted:1 plan:1 special:2 airport:1 initialize:10 equal:7 once:3 runtimes:2 identical:2 report:2 intelligent:2 few:2 randomly:3 ve:4 lite:1 recalculates:1 atlanta:2 organization:1 navigation:1 bracket:1 edge:17 respective:1 orthogonal:1 tree:9 old:3 desired:1 increased:1 obstacle:5 cost:19 vertex:91 front:1 reported:1 density:3 international:1 contract:1 together:1 manage:2 priority:30 return:9 reusing:1 north:1 matter:1 satisfy:2 performed:1 view:1 break:2 traffic:1 reached:3 start:25 maintains:4 became:3 variance:1 skoenig:1 correspond:5 cc:1 researcher:2 sebastian:1 ed:1 proof:2 stop:1 knowledge:3 back:4 wesley:1 originally:2 spaccamela:1 furthermore:1 just:2 until:6 koenig:2 hand:1 web:1 gridworlds:4 quality:2 gray:3 artifact:1 grows:1 b3:1 concept:1 read:1 during:6 replan:3 sponsoring:1 leftmost:1 ramalingam:1 complete:3 demonstrate:2 performs:1 percent:4 common:1 pseudocode:2 qp:2 overview:2 kluwer:1 refer:1 blocked:6 grid:3 consistency:2 similarly:2 hp:4 moving:4 access:1 longer:1 add:1 optimizing:1 route:5 binary:1 arbitrarily:2 rep:1 ced:1 minimum:2 determine:3 shortest:19 ii:1 branch:1 technical:1 faster:5 academic:1 award:1 a1:1 va:4 baa:1 repetitive:1 sometimes:1 cell:25 want:1 separately:1 decreased:1 interval:1 else:5 publisher:1 fifty:3 sure:1 inconsistent:6 call:4 easy:1 concerned:1 pratt:1 affect:1 topology:2 idea:1 itr:1 knowing:2 whether:1 effort:2 likhachev:2 queue:22 sgoal:10 locally:18 nsf:1 notice:1 sstart:10 affected:3 group:1 recomputed:4 four:5 key:13 deletes:1 d3:1 changing:2 breadth:7 graph:8 underconsistent:7 sum:1 fourth:1 decide:1 decision:3 guaranteed:1 display:1 lpa:42 infinity:3 stentz:2 expanded:2 speedup:1 developing:1 according:4 smaller:3 slightly:1 terminates:2 making:2 repair:1 equation:3 turn:2 count:1 addison:1 operation:3 apply:1 eight:10 encounter:1 original:7 top:3 denotes:4 a4:1 maintaining:1 build:1 implied:1 already:1 strategy:1 traditional:1 diagonal:1 distance:25 thank:1 thrun:2 restating:1 reason:1 assuming:1 length:1 relationship:1 potentially:3 trace:4 stated:2 marchetti:1 implementation:2 policy:1 unknown:1 upper:1 observation:1 finite:3 situation:4 precise:1 gridworld:16 namely:3 optimized:2 pop:2 pearl:1 rut:1 able:1 fp:11 biasing:3 overconsistent:10 including:1 representing:1 technology:2 carried:1 kj:1 literature:3 determining:1 manhattan:1 fully:1 analogy:1 sufficient:1 consistent:11 editor:1 share:1 row:1 changed:24 supported:1 allow:1 institute:2 lifelong:15 focussed:2 absolute:1 calculated:1 maze:3 computes:4 author:1 c5:1 forward:1 made:2 approximate:1 forever:2 status:2 incoming:1 assumed:1 unnecessary:1 consuming:1 percolate:1 search:84 table:2 learn:1 expanding:1 expansion:5 investigated:1 complex:1 anthony:1 domain:1 official:1 main:4 rh:16 backup:1 nothing:1 child:1 georgia:3 position:1 breaking:1 third:2 theorem:1 remained:2 sequential:2 iel:1 simply:1 expressed:1 contained:1 applies:2 dynamicswsf:11 corresponds:2 determines:2 goal:11 towards:1 change:13 experimentally:1 called:3 total:4 partly:1 experimental:5 college:2 support:1 calculatekey:7 scratch:4
1,101
2,004
Scaling laws and local minima in Hebbian ICA Magnus Rattray and Gleb Basalyga Department of Computer Science, University of Manchester, Manchester M13 9PL, UK. magnus@cs.man.ac.uk, basalygg@cs.man.ac.uk Abstract We study the dynamics of a Hebbian ICA algorithm extracting a single non-Gaussian component from a high-dimensional Gaussian background. For both on-line and batch learning we find that a surprisingly large number of examples are required to avoid trapping in a sub-optimal state close to the initial conditions. To extract a skewed signal at least examples are required for -dimensional data and examples are required to extract a symmetrical signal with non-zero kurtosis.      1 Introduction Independent component analysis (ICA) is a statistical modelling technique which has attracted a significant amount of research interest in recent years (for a review, see Hyv?arinen, 1999). The goal of ICA is to find a representation of data in terms of a combination of statistically independent variables. A number of neural learning algorithms have been applied to this problem, as detailed in the aforementioned review. Theoretical studies of ICA algorithms have mainly focussed on asymptotic stability and efficiency, using the established results of stochastic approximation theory. However, in practice the transient stages of learning will often be more significant in determining the success of an algorithm. In this paper a Hebbian ICA algorithm is analysed in both on-line and batch mode, highlighting the critical importance of the transient dynamics. We find that a surprisingly large number of training examples are required in order to avoid trapping in a sub-optimal state close to the initial conditions. To detect a skewed signal at least examples are required for -dimensional data, while examples are required for a symmetric signal with non-zero kurtosis. In addition, for on-line learning we show that the maximal initial learning rate which allows successful learning is unusually low, being for a skewed signal and for a symmetric signal.            In order to obtain a tractable model, we consider the limit of high-dimensional data and study an idealised data set in which a single non-Gaussian source is mixed into a large number of Gaussian sources. Recently, one of us considered a more general model in which an arbitrary, but relatively small, number of non-Gaussian sources were mixed into a high-dimensional Gaussian background (Rattray, 2002). In that work a solution to the dynamics of the on-line algorithm was obtained in closed form for learning iterations and a simple solution to the asymptotic dynamics under the optimal learning rate decay was obtained. However, it was noted there that modelling the dynamics on an timescale is not always appropriate, because the algorithm typically requires much longer in order to    escape from a class of metastable states close to the initial conditions. In order to elucidate this effect in greater detail we focus here on the simplest case of a single non-Gaussian source and we will limit our analysis to the dynamics close to the initial conditions. In recent years a number of on-line learning algorithms, including back-propagation and Sanger?s PCA algorithm, have been studied using techniques from statistical mechanics (see, for example, Biehl (1994); Biehl and Schwarze (1995); Saad and Solla (1995) and contributions in Saad (1998)). These analyses exploited the ?self-averaging? property of certain macroscopic variables in order to obtain ordinary differential equations describing the deterministic evolution of these quantities over time in the large limit. In the present case the appropriate macroscopic quantity does not self-average and fluctuations have to be considered even in the limit. In this case it is more natural to model the on-line learning dynamics as a diffusion process (see, for example Gardiner, 1985).  2 Data Model In order to apply the Hebbian ICA algorithm we must first sphere the data, ie. linearly transform the data so that it has zero mean and an identity covariance matrix. This can be achieved by standard transformations in a batch setting or for on-line learning an adaptive sphering algorithm, such as the one introduced by Cardoso and Laheld (1996), could be used. To simplify the analysis it is assumed here that the data has already been sphered. Without loss of generality it can also be assumed that the sources each have unit variance.          Each data point is generated from a noiseless linear mixture of sources which are decomposed into a single non-Gaussian source and uncorrelated Gaussian components, . We will also decompose the mixing matrix into a column vector and a rectangular matrix associated with the non-Gaussian and Gaussian components respectively,      !   #"$% '& *  )( (1) is presented to the We will consider both the on-line case, in which a new IID example algorithm at each time and then discarded, and also the batch case, in which a finite set of examples are available to the algorithm. To conform with the model assumptions the mixing matrix must be unitary, which leads to the following constraints,  + ,- /.  10  ,2 0 "3   0  4   0 + %0  1 0   10  % 7 & 50    .  %0 ,6%0     (2) (3) 3 On-line learning B<:'8%0CC  8 such that the projection 9;:<8=013>@?A . Defining 9  8 0   #"D     (4)  BE#"GFIH JKJ 8LJMJ B where F   NO P  The goal of ICA is to find a vector the overlap we obtain, 8 RB >S?  8 where we have made use of the constraint in eqn. (2). This assumes zero correlation between and which is true for on-line learning but is only strictly true for the first iteration of batch learning (see section 4). In the algorithm described below we impose a normalisation constraint on such that . In this case we see that the goal is to find such that . JMJ 8LJKJQ  8 A simple Hebbian (or anti-Hebbian) learning rule was studied by Hyv?arinen and Oja (1998), who showed it to have a remarkably simple stability condition. We will consider the deflationary form in which a single source is learned at one time. The algorithm is closely related to Projection Pursuit algorithms, which seek interesting projections in highdimensional data. A typical criteria for an interesting projection is to find one which is maximally non-Gaussian in some sense. Maximising some such measure (simple examples would be skewness or kurtosis) leads to the following simple algorithm (see Hyv?arinen and Oja, 1998, for details). The change in at time is given by, 8 *  & (5) ( followed by normalisation such that JKJ 8LJKJ   Here,  is the learning rate and  9 is some non-linear function  which  we will take to be at least three times differentiable. An even non-linearity, eg.  9  9 , is appropriate for detecting asymmetric signals while a more common choice is an odd function, eg.   9   9 or   9      9 E, which used to detect symmetric non-Gaussian P  hascanto bebe chosen signals. In the latter case  in order to ensure stability of the correct solution, as described by Hyv? a rinen and Oja (1998), either adaptively or using a?  in the case of an even non-linearity. Remarkably, the same priori knowledge. We set   non-linearity can be used to separate both sub and super-Gaussian signals, in contrast to 8  9 ( maximum likelihood methods for which this is typically not the case. We can write the above algorithm as,   (6) 8 (   H  "8% (  9"(  9 (  "9/ (   ,(  9 (  JMJ ( JMJ  &  and    (two different scalings will be considered below) we can For large expand out to get a weight decay normalisation, 8 (   8 ( "  9 ( ! (  9 ( 8 (#"  $       9 (  8 ( & (7) Taking the dot-product with   gives the following update increment for the overlap B , % B<&  9   (  B ( 9 ( "         9 (  B ( (8) where we used %the constraint in eqn. (3) to set ,0   . Below we calculate the mean and variance of B for two different scalings of the learning rate. Because the conditional in eqn. 4) these expressions distribution for 9 given  only depends on B (setting JKJ 8LJKJ will depend only on B and statistics of the non-Gaussian source distribution. 3.1 Dynamics close to the initial conditions , 8  B<   '  If the entries in and are initially of similar order then one would expect . This is the typical case if we consider a random and uncorrelated choice for and the initial entries in . Larger initial values of could only be obtained with some prior knowledge of the mixing matrix which we will not assume. We will set in the following discussion, where is assumed to be an quantity. The discussion below is therefore restricted to describing the dynamics close to the initial conditions. For an account of the transient dynamics far from the initial conditions and the asymptotic dynamics close to an optimal solution, see Rattray (2002). 8 3.1.1 B ( (;: B*)      9  even, + - , N  9   9  If the signal is asymmetrical then an even non-linearity can be used, for example is a common choice. In this case the appropriate (ie. maximal) scaling for the learning rate is and we set where is an scaled learning rate parameter. In      /.10   .         even,    %    odd,    %  (: 8=01  )     Figure 1: Close to the initial conditions (where ) the learning dynamics is equivalent to diffusion in a polynomial potential. For asymmetrical source distributions we can use an even non-linearity in which case the potential is cubic, as shown on the left. For symmetrical source distributions with non-zero kurtosis we should use an odd non-linearity in which case the potential is quartic, as shown on the right.  The  . with a potential barrier dynamics is initially confined in a metastable state near (E N % ( at each iteration are given + %        F  .  ( "  +     F  .(  "    E ( . (9)       + %      (10) Var ( .  F .   this case we find that the mean and variance of the change in by (to leading order in ), + +% (   .  F   ON  where is the third cumulant of the source distribution (third central moment), which measures skewness, and brackets denote averages over . We also find that  E for integer ! . In this case the system can be described by a Fokker-Planck equation for large (see, for example, Gardiner, 1985) with a characteristic timescale of . The system is locally equivalent to a diffusion in the following cubic potential,  "   #     (11)                F   .   (    +   F  .( with a diffusion coefficient $ %    F . which is independent% ofmust ( . The shape of this be overcome to potential is shown on the left of fig. 1. A potential barrier of (  escape a metastable state close to the initial conditions. 3.1.2   9  odd, + "  , N   9   9   9      9      5  . 0  If the signal is symmetrical, or only weakly asymmetrical, it will be necessary to use an odd non-linearity, for example or are popular choices. In this case a lower learning rate is required in order to achieve successful separation. The appropriate scaling for the learning rate is and we set where again is an scaled learning rate parameter. In this case we find that the mean and variance of the change in at each iteration are given by,   ( . + % ( .        F  .  ( " # + "       F  .( "   (12)      + %       F . Var ( . (13) where + " is the fourth cumulant source distribution (measuring kurtosis) and brackets  NO Pof  .theAgain denote averages over F  the system can be described by a Fokker-Planck E     equation for large but in this case the timescale for learning is , an order of slower than in the asymmetrical case. The system is locally equivalent to diffusion in the following quartic potential,  "     F    .   (      " J + "      F  J .( "   (14) with a diffusion coefficient $    F . . We have assumed < Sign + " which is a necessary condition for successful learning. In the case of a cubic non-linearity this is  (  also the condition for stability of the optimal fixed point, although in general these two conditions may not be equivalent (Rattray, 2002). The of this potential is shown  shape on the right of fig. 1 and again a potential barrier of must be overcome to escape a metastable state close to the initial conditions. % B< N For large . the dynamics of ( corresponds to an Ornstein-Uhlenbeck process with a Gaussian stationary distribution of fixed unit variance. Thus, if one chooses too large . initially N (recall, B  (0 )  ). As . is reduced the dynamics will become localised close to B  3.1.3 Escape times from a metastable state at . the potential barrier confining the dynamics is reduced. The timescale for escape for large (mean first passage time) is mainly determined by the effective size of the barrier (see, for example, Gardiner, 1985), % where  *     escape   % $ (15) * is a unit of time in   N , + .5:    . for even  9 , +  ,   N . + .%:&   . (16) for odd  9 , + "  , is the potential barrier, $ is the diffusion coefficient and the diffusion process. For the two cases considered above we obtain,   even escape  odd escape     F  .      +      F       F  .  J + "       F  J         The constants of proportionality depend on the shape of the potential and not on . As the learning rate parameter is reduced so the timescale for escape is also reduced. However, the choice of optimal learning rate is non-trivial and cannot be determined by considering only the leading order terms in as above, because although small will result in a quicker , this in turn will lead to a very slow escape from the unstable fixed point near learning transient after escape. Notice that escape time is shortest when the cumulants or " are large, suggesting that deflationary ICA algorithms will tend to find these signals first. B . B  N + +    From the above discussion one can draw two important conclusions. Firstly, the initial learning rate must be less than initially in order to avoid trapping close to the initial conditions. Secondly, the number of iterations required to escape the initial transient will be greater than , resulting in an extremely slow initial stage of learning for large . The most extreme case is for symmetric source distributions with non-zero kurtosis, in which case learning iterations are required.     In fig. 2 we show results of learning with an asymmetric source (top) and uniform source (bottom) for different scaled learning rates. As the learning rate is increased (left to right) we observe that the dynamics becomes increasingly stochastic, with the potential barrier becoming increasingly significant (potential maxima are shown as dashed lines). For the largest value of learning rate (  ) the algorithm becomes trapped close to the initial conditions for the whole simulation time. From the time axis we observe that the learning timescale is for the asymmetrical signal and for the symmetric signal, as predicted by our theory.   .$   ?=0.1 1 ?=1 ?=5 1 1 R R 0.5 0.5 0.5 0 0 0 2 ?(x)=x , ? ? 0 R 3 0 5 10 t/N2 15 1 R ?(x)=x3, ? ? 0 4 0.5 0 5 10 t/N2 15 0 1 R 0.5 1 R 0.5 0 0 0 ?0.5 ?0.5 ?0.5 ?1 0 5 t/N3 +  & 10 ?1 0    NQN 5 t/N3 10 5 ?1 0 10 t/N2 15 5 t/N3 10 Figure 2: 100-dimensional data ( ) is produced from a mixture containing a single non-Gaussian source. In the top row we show results for a binary, asymmetrical source with skewness . In the bottom row we show results for a uniformly  and  distributed source and . Each row shows learning with the same initial conditions and data but with different scaled learning rates (left to right and  ) where (top) or (bottom). Dashed lines are maxima of the potentials in fig. 1. . :    .         :  .R N & Q  4 Batch learning The batch version of eqn. (5) for sufficiently small learning rates can be written, % 8      9 (  (  9 ( 8 ( " (   (17)  where  is the number of training examples. Here we argue that such an update requires at least the same order of examples as in the on-line case, in order to be successful. Less data will result in a low signal-to-noise ratio initially and the possibility of trapping in a sub-optimal fixed point close to the initial conditions. As in the on-line case we can write the update in terms of  B B , % B      9 (   (  9 ( B ( " & (  N (18)  We make an assumption that successful learning is unlikely unless the initial increment in is in the desired direction. For example, with an asymmetric signal and quadratic nonlinearity we require initially, while for a symmetric signal and odd non-linearity we require % . We have carried out simulations of batch learning which confirm that a relatively low percentage of runs in which the intial increment was incorrect result in successful learning compared to typical performance. As in the on-line case we observe that runs either succeed, in which case , or fail badly with remaining . % %B B + N B    '   B > ? B  '  initially and we can therefore expand the right-hand side of As before, B  %  % eqn. (18) in orders of B for large . B ( B at the first iteration) is a sum over raninit %B domly sampled terms, and the central limit theorem states that for large  the distribution init from which is sampled will be Gaussian, with mean and variance given by (to leading order in ), B % B      +     F  B  " # + "       F  B  (19)    %     &  Var B (20)    F Notice that the + term disappears in the case of an asymmetrical non-linearity, which is why we have left both % terms in eqn. (19). The algorithm will be likely to fail when is of the same order (or greater) than the mean. Since the standard deviation of B B  '  initially, we see that this is true for R     in the case of an even nonE init  init  init R     linearity and asymmetric signal, or for  in the case of an odd non-linearity and a signal with non-zero kurtosis. We expect these results to be necessary but not necessarily sufficient for successful learning, since we have only shown that this order of examples is the minimum required to avoid a low signal-to-noise ratio in the first learning iteration. A complete treatment of the batch learning problem would require much more sophisticated formulations such as the mean-field theory of Wong et al. (2000). 5 Conclusions and future work In both the batch and on-line Hebbian ICA algorithm we find that a surprisingly large number of examples are required to avoid a sub-optimal fixed point close to the initial conditions. We expect simialr scaling laws to apply in the case of small numbers of non-Gaussian sources. Analysis of the square demixing problem appears to be much more challenging as in this case there may be no simple macroscopic description of the system for large . It is therefore unclear at present whether ICA algorithms based on Maximum-likelihood and Information-theoretic principles (see, for example, Bell and Sejnowski, 1995; Amari et al., 1996; Cardoso and Laheld, 1996), which estimate a square demixing matrix, exhibit similar classes of fixed point to those studied here.  Acknowledgements: This work was supported by an EPSRC award (ref. GR/M48123). We would like to thank Jon Shapiro for useful comments on a preliminary version of this paper. References S-I Amari, A Cichocki, and H H Yang. In D S Touretzky, M C Mozer, and M E Hasselmo, editors, Neural Information Processing Systems 8, pages 757?763. MIT Press, Cambridge MA, 1996. A J Bell and T J Sejnowski. Neural Computation, 7:1129?1159, 1995. M Biehl. Europhys. Lett., 25:391?396, 1994. M Biehl and H Schwarze. J. Phys. A, 28:643?656, 1995. J-F Cardoso and B. Laheld. IEEE Trans. on Signal Processing, 44:3017?3030, 1996. C. W. Gardiner. Handbook of Stochastic Methods. Springer-Verlag, New York, 1985. A Hyv?arinen. Neural Computing Surveys, 2:94?128, 1999. A Hyv?arinen and E Oja. Signal Processing, 64:301?313, 1998. M Rattray. Neural Computation, 14, 2002 (in press). D Saad, editor. On-line Learning in Neural Networks. Cambridge University Press, 1998. D Saad and S A Solla. Phys. Rev. Lett., 74:4337?4340, 1995. K Y M Wong, S Li, and P Luo. In S A Solla, T K Leen, and K-R M?uller, editors, Neural Information Processing Systems 12. MIT Press, Cambridge MA, 2000.
2004 |@word version:2 polynomial:1 proportionality:1 hyv:6 gfih:1 seek:1 simulation:2 covariance:1 moment:1 initial:22 luo:1 analysed:1 attracted:1 must:4 written:1 shape:3 update:3 stationary:1 trapping:4 deflationary:2 detecting:1 jkj:3 firstly:1 differential:1 become:1 incorrect:1 ica:11 mechanic:1 decomposed:1 considering:1 becomes:2 pof:1 linearity:12 skewness:3 transformation:1 unusually:1 scaled:4 uk:3 unit:3 planck:2 before:1 local:1 limit:5 fluctuation:1 becoming:1 studied:3 challenging:1 statistically:1 practice:1 x3:1 laheld:3 bell:2 projection:4 get:1 cannot:1 close:15 wong:2 equivalent:4 deterministic:1 rectangular:1 survey:1 rule:1 stability:4 increment:3 elucidate:1 rinen:1 asymmetric:4 bottom:3 epsrc:1 quicker:1 calculate:1 solla:3 mozer:1 dynamic:17 depend:2 weakly:1 efficiency:1 effective:1 sejnowski:2 europhys:1 larger:1 biehl:4 amari:2 statistic:1 gleb:1 timescale:6 transform:1 differentiable:1 kurtosis:7 maximal:2 product:1 mixing:3 achieve:1 description:1 manchester:2 ac:2 odd:9 c:2 predicted:1 direction:1 closely:1 correct:1 stochastic:3 transient:5 require:3 arinen:5 decompose:1 preliminary:1 secondly:1 strictly:1 pl:1 sufficiently:1 considered:4 magnus:2 domly:1 largest:1 hasselmo:1 uller:1 mit:2 gaussian:19 always:1 super:1 avoid:5 focus:1 modelling:2 likelihood:2 mainly:2 contrast:1 detect:2 sense:1 typically:2 unlikely:1 initially:8 expand:2 aforementioned:1 priori:1 field:1 jon:1 future:1 simplify:1 escape:13 oja:4 interest:1 normalisation:3 possibility:1 mixture:2 bracket:2 extreme:1 necessary:3 unless:1 desired:1 theoretical:1 increased:1 column:1 cumulants:1 measuring:1 ordinary:1 deviation:1 entry:2 intial:1 uniform:1 successful:7 gr:1 too:1 chooses:1 adaptively:1 ie:2 again:2 central:2 containing:1 leading:3 li:1 account:1 potential:16 suggesting:1 coefficient:3 depends:1 ornstein:1 closed:1 contribution:1 square:2 variance:6 who:1 characteristic:1 produced:1 iid:1 none:1 phys:2 touretzky:1 associated:1 sampled:2 treatment:1 popular:1 recall:1 knowledge:2 sophisticated:1 back:1 appears:1 maximally:1 formulation:1 leen:1 generality:1 stage:2 correlation:1 hand:1 eqn:6 propagation:1 mode:1 schwarze:2 effect:1 true:3 asymmetrical:7 idealised:1 evolution:1 symmetric:6 eg:2 skewed:3 self:2 noted:1 criterion:1 complete:1 theoretic:1 passage:1 recently:1 common:2 significant:3 cambridge:3 nonlinearity:1 dot:1 longer:1 recent:2 showed:1 quartic:2 certain:1 verlag:1 binary:1 success:1 exploited:1 minimum:2 greater:3 impose:1 shortest:1 signal:22 dashed:2 hebbian:7 sphere:1 award:1 noiseless:1 iteration:8 uhlenbeck:1 achieved:1 confined:1 background:2 remarkably:2 addition:1 source:20 macroscopic:3 saad:4 comment:1 tend:1 integer:1 extracting:1 unitary:1 near:2 yang:1 whether:1 expression:1 pca:1 york:1 useful:1 detailed:1 cardoso:3 amount:1 locally:2 simplest:1 reduced:4 shapiro:1 percentage:1 notice:2 sign:1 trapped:1 rb:1 rattray:5 conform:1 write:2 diffusion:8 year:2 sum:1 run:2 fourth:1 separation:1 draw:1 scaling:6 followed:1 quadratic:1 badly:1 gardiner:4 constraint:4 n3:3 extremely:1 relatively:2 sphering:1 department:1 metastable:5 combination:1 increasingly:2 rev:1 restricted:1 equation:3 describing:2 turn:1 fail:2 tractable:1 available:1 pursuit:1 apply:2 observe:3 appropriate:5 batch:10 slower:1 assumes:1 top:3 ensure:1 remaining:1 sanger:1 already:1 quantity:3 unclear:1 exhibit:1 separate:1 thank:1 argue:1 unstable:1 trivial:1 maximising:1 ratio:2 basalyga:1 localised:1 confining:1 discarded:1 finite:1 anti:1 defining:1 arbitrary:1 introduced:1 required:11 learned:1 established:1 trans:1 below:4 including:1 critical:1 overlap:2 natural:1 disappears:1 axis:1 carried:1 extract:2 cichocki:1 review:2 prior:1 acknowledgement:1 determining:1 asymptotic:3 law:2 loss:1 expect:3 mixed:2 interesting:2 var:3 sufficient:1 principle:1 editor:3 uncorrelated:2 row:3 surprisingly:3 supported:1 side:1 focussed:1 taking:1 barrier:7 distributed:1 overcome:2 lett:2 made:1 adaptive:1 far:1 confirm:1 handbook:1 symmetrical:3 assumed:4 why:1 init:4 m13:1 necessarily:1 linearly:1 whole:1 noise:2 n2:3 ref:1 fig:4 cubic:3 slow:2 sub:5 third:2 theorem:1 decay:2 demixing:2 importance:1 likely:1 jmj:3 highlighting:1 springer:1 fokker:2 corresponds:1 ma:2 succeed:1 conditional:1 goal:3 identity:1 man:2 change:3 typical:3 determined:2 uniformly:1 averaging:1 highdimensional:1 latter:1 sphered:1 cumulant:2
1,102
2,005
Audio-Visual Sound Separation Via Hidden Markov Models John Hershey Department of Cognitive Science University of California San Diego Michael Casey Mitsubishi Electric Research Labs Cambridge, Massachussets jhershey@cogsci.ucsd.edu casey@merl.com Abstract It is well known that under noisy conditions we can hear speech much more clearly when we read the speaker's lips. This suggests the utility of audio-visual information for the task of speech enhancement. We propose a method to exploit audio-visual cues to enable speech separation under non-stationary noise and with a single microphone. We revise and extend HMM-based speech enhancement techniques, in which signal and noise models are factori ally combined, to incorporate visual lip information and employ novel signal HMMs in which the dynamics of narrow-band and wide band components are factorial. We avoid the combinatorial explosion in the factorial model by using a simple approximate inference technique to quickly estimate the clean signals in a mixture. We present a preliminary evaluation of this approach using a small-vocabulary audio-visual database, showing promising improvements in machine intelligibility for speech enhanced using audio and visual information. 1 Introduction We often take for granted the ease with which we can carryon a conversation in the proverbial cocktail party scenario: guests chatter, glasses clink, music plays in the background: the room is filled with ambient sound. The vibrations from different sources and their reverberations coalesce translucently yielding a single time series at each ear, in which sounds largely overlap even in the frequency domain. Remarkably the human auditory system delivers high-quality impressions of sounds in conditions that perplex our best computational systems. A variety of strategies appear to be at work in this, including binaural spatial analysis, and inference using prior knowledge of likely signals and their contexts. In speech perception, vision often plays a crucial role, because we can follow in the lips and face the very mechanisms that modulate the sound, even when the sound is obscured by acoustic noise. It has been demonstrated that the addition of visual cues can enhance speech recog- nition as much as removing 15 dB of noise [1]. Vision provides speech cues that are complementary to audio cues such as components of consonants and vowels that are likely to be obscured by acoustic noise [2]. Visual information is demonstra- bly beneficial to HMM-based automatic speech recognition (ASR) systems, which typically suffer tremendously under moderate acoustical noise [3]. We introduce a method of audio-visual speech enhancement using factorial hidden Markov models (fHMMs). We focus on speech enhancement rather than speech recognition for two reasons: first, speech conveys useful paralinguistic information, such as prosody, emotion, and speaker identity, and second, speech contains useful cues for separation from noise, such as pitch. In automatic speech recognition (ASR) systems, these cues are typically discarded in an effort to reduce irrelevant variance among speakers and utterances within a phonetic class. Whereas the benefit of vision to speech recognition is well known, we may well wonder if visual input offers similar benefits to speech enhancement. In [4] a nonparametric density estimator was used to adapt audio and video transforms to maximize the mutual information between the face of a target speaker and an audio mixture containing both the target voice and a distracter voice. These transforms were then used to construct a stationary filter for separating the target voice from the mixture without any prior knowledge or training. In [5] a multi-layer perceptron is trained to map noisy estimates of formants to clean ones, employing lip parameters (width, height and area of the lip opening) extracted from video as additional input. The re-estimated formant contours were used to filter the speech to enhance the signal. In both cases video information improved signal separation. Neither system, however, made use of the dynamics of speech. In speech recognition, HMMs are commonly used because of the advantages of modeling signal dynamics. This suggests the following strategy: train an audiovisual HMM on clean speech, infer the likelihoods of its state sequences, and use the inferred state probabilities of the signal and noise to estimate a sequence of filters to clean the data. In cases where background noise also has regularity, such as the combination of two voices, another HMM can be used to model the background noise. Ephraim [6] first proposed an approach to factorially combining two HMMs in such an enhancement system. In [7] an efficient variational learning rule for the factorial HMM is formulated, and in [8, 9] fHMM speech enhancement was recently revived using some clever tricks to allow more complex models. The fHMM approach is amenable to audio-visual speech enhancement in many different forms. In the simplest formulation, which we pursue here, the signal observation model includes visual features. These visual inputs constrain the signal HMM and produce more accurate filters. Below we present a prototype architecture for such a system along with preliminary results. 1 1.1 Factorial Speech Models One of the challenges of using speech HMMs for enhancement is to model speech in sufficient detail. Typically, speech models, following the practice in ASR, ignore narrow-band, spectral details (corresponding to upper cepstral components) which carry pitch information, because they tend to vary across speakers and utterances for the same word or phoneme. Instead such systems focus on the smooth, or wideband, spectral characteristics (corresponding to lower cepstral components) such as are produced by the articulation of the mouth. Such wide-band spectral patterns loosely represent formant patterns, a well-known cue for vowel discrimination. In cases where the pitch or other narrow-band properties, of the background signals differ from the foreground speech, and have predictable dynamics, such as with lWe defer a detailed mathematical development to subsequent publications. Contact jhershey@cogsci.ucsd.edu for further information two simultaneous speech signals, these components may be helpful in separating the two signals. Figure 1 illustrates the analysis of two words into wide-band and narrow-band components. "one" " two" Full band: Narrow band: Wide band: Figure 1: full-band, narrow-band, and wide-band log spectrograms of two words. The wide-band log spectrograms (bottom) are derived by low-pass filtering the log spectra (across the frequency domain), and the narrow-band log spectrograms (middle) derived by high pass filtering the log spectra The full log spectrogram (top) is the sum of the two. However, the wide-band and narrow-band variations in speech are only loosely coupled. For instance, a given formant is likely to be uttered with many different pitches and a given pitch may be used to utter any formant. Thus a model of the full spectrum of speech would have to have enough states to represent every combination of pitches and formants. Such a model requires a large amount of training data and imposes serious computational burdens. For instance in [8] a model with 8000 states is employed. When combined with a similarly complex noise model, the composite model has 64 million states. This is expensive in terms of computation as well as the number of data points required for inference. To parsimoniously model the complexity of speech, we employ a factorial HMM for a single speech signal, in which wide and narrow-band components are represented in sub-models with independent dynamics. We therefore train the two submodels independently using Gaussian observation probability density functions (p.d.f.) on the wide-band or narrow-band log spectra, with diagonal covariances for the sake of simplicity. Figure 2(a) depicts the graphical model for a single wide or narrow-band component. Narrow-Band Slate Di screte States Wide- Band Stale Continuou s Observation s Combined Observmions (a) simple HMM (b) factorial speech HMM Figure 2: single HMMs are trained separately on wide-band and narrow-band speech signals (a) and then combined factorially in (b) by adding the means and variances of their observation distributions To combine the sub-models, we have to specify the observation p.d.f. for a combination of a wide and a narrow-band state, over the log-spectrum of speech prior to liftering. Because the observation densities of each component are Gaussian, and the log-spectra of the wide and narrow-band components add in the log spectrum, the composite state has a Gaussian observation p.d.f., whose mean and variance is the sum of the component observation means and variances. Although the states of the two sub-models are marginally independent they are typically conditionally dependent given the observation sequence. In other words we assume that the state dependencies between the sub-models for a given speech signal can be explained entirely via the observations. Figure 2(b) depicts the combination of the wide and narrow-band models, where the observation p.d.f. 's are a function of two state variables. When combining the signal and noise models (or two different speech models) in contrast, the signals add in the frequency domain, and hence in the log spectral domain they longer simply add. In the spectral domain the amplitudes of the two signals have log-normal distributions, and the relative phases are unknown. There is no closed form distribution for the sum of two random variables with log-normal amplitudes and a uniformly distributed phase difference. Disregarding phase differences we apply a well-known approximation to the sum of two lognormal random variables, in which we match the mean and variance of a lognormal random variable to the sum of the means and variances of the two component lognormal random variables [10]. Phase uncertainty can also be incorporated into an approximation; however in practice the costs appear to outweigh the benefits.2 Figure 3(a) depicts the combination of two factorial speech models, where the observation p.d.f.s are a function of two state variables. -'-- .. . Video Observations ~ Audio Observations (a) dual factorial HMM 6 (1f 0 (b) speech fHMM with video Figure 3: combining two speech fHMMs (a) and adding video observations to a speech fHMM (b). Using the log-normal observation distribution of the composite model we can estimate the likelihood of the speech and noise states for each frame using the well known forward-backward recursion. For each frame of the test data we can compute the expected value of the amplitude of each model in each frequency bin. Taking 2The uncertainty of the phase differences can be incorporated by modeling the sum as a mixture of lognormals that uniformly samples phase differences. Each mixture element is approximated by taking as its mean the length of the sum of the mean amplitudes when added in the complex plane according a particular phase difference, and as its variance the sum of the two variances. This estimation is facilitated by the assumption of diagonal covariances in the log spectral domain. the expected value of the signal in the numerator and the expected value of the signal plus noise in the denominator yields a Wiener filter which is applied to the original noisy signal enhancing the desired component. When we have two speech signals one person's noise is another's signal and we can separate both by the same method. 2 Incorporating vision We incorporate vision after training the audio models in order to test the improvement yielded by visual input while holding the audio model constant. A video observation distribution is added to each state in the model by obtaining the probability of each state in each frame of the audio training data using the forward-backward procedure, then estimating the parameters of the video observation distributions for each state, in the manner of the Baum-Welch observation re-estimation formula. This procedure is iterated until it converges. In this way we construct a system in which the visual observations are modular. Figure 3(b) depicts the structure ofthe resulting speech model. Such a method in which audio and visual features are integrated early in processing is only one of several approaches. We envision other late integration approaches in which audio and visual dynamics are more loosely coupled. What method of audio-visual integration may be best for this task is an open question. 3 Efficient inference In the models described above, in which we factorially combine two speech models , each of which is itself factorial , the complexity of inference in the composite model, using the forward-backward recursion, can easily become unmanageable. If K is the number of states in each subcomponent, then K4 is the number of states in the composite HMM. In our experiments K is on the order of 40 states, so there are 2,560,000 states in the composite model. Naively each composite state must be searched when computing the probabilities of state sequences necessary for inference. Interesting approximation schemes for similar models are developed in [8, 9]. We develop an approximation as follows. Rather than computing the forward-backward procedure on the composite HMM, we compute it sequentially on each sub-HMM to derive the probability of each state in each frame. Of course, in order to evaluate the observation probabilities of the current sub-HMMs for a given frame, we need to consider the state probabilities of the other three sub-HMMs, because their means and variances are combined in the observation model. These state probabilities and their associated observation probabilities comprise a mixture model for a given frame. The composite mixture model still has K4 states, so to defray this complexity during forward-backward analysis of the current sub-HMM, for each frame we approximate the observation mixtures of each of the other three sub-HMMs with a single Gaussian, whose mean and variance matches that of the mixture. Thus we only have to consider the K states of the current model, and use the summarized means and variances of the other three HMMs as auxiliary inputs to the observation model. We initialize the state probabilities in each frame with the equilibrium distribution for each sub-HMM. In our experiments, after a handful of iterations, the composite state probabilities tend to converge. This method is closely related to a structured variational approximation for factorial HMMs [7] and can be also be seen as an approximate belief propagation or sum-product algorithm [11]. 4 Data We used a small-vocabulary audio-visual speech database developed by Fu Jie Huang at Carnegie Mellon University 3 [12]. These data consist of audio and video recordings of 10 subjects (7 males and 3 females) saying 78 isolated words commonly used for numbers and time, such aS,"one" "Monday", "February", "night", etc. The sequence of 78 words is repeated in 10 different takes. Half of these takes were used for training, and one of the remaining takes was used as the test set. The data set included outer lip parameters extracted from video using an automatic lip tracker, including height of the upper and lower lips relative to the corners the width from corner to corner. We interpolated these lip parameters to match the audio frame rate, and calculate time derivatives. Audio consisted of 16-bit, 44.1 kHz recordings which we resample to 8000 kHz. The audio was framed at 60 frames per second, with an overlap of 50% , yielding 264 samples per frame. 4 The frames were analyzed into cepstra: the wide-band log spectrum is derived from the lower 20 cepstral components and the wide-band log spectrum from the upper cepstra. 5 Results Speaker dependent wide and narrow-band HMMs having 40 states each were trained on data from two subjects (" Anne" and" Chris") selected from the training set. A PCA basis was used to reduce the log spectrograms to a more manageable size of 30 dimensions during training. This resulted in some non-zero covariances near the diagonal in the learned observation covariance matrices, which we discarded. An entropic prior and parameter extinction were used to sparsify the transition matrices during training [13]. The narrow-band model learned states that represented different pitches and had transition probabilities that were non-zero mainly between neighboring pitches. The narrow-band model's video observation probability distributions were largely overlapping, reflecting the fact that video tells us little about pitch. The wide-band model learned states that represented different formant structures. The video observation distributions for several states in the wide-band model were clearly separated, reflecting the information that video provides about the formant structure. Subjectively the enhanced signals sound well separated from each other for the most part. Figure 4(a) (bottom) shows the estimated spectrograms for a mixture of two different words spoken by the same speaker - an extremely difficult task. To quantify these results we evaluate the system using speech recognizer, on the slightly easier task of separating the speech of the two different speakers, whose voices were in different but overlapping pitch ranges. A test set was generated by mixing together 39 randomly chosen pairs of words, one from each subject, such that no word was used twice. Each word pair was mixed at five different signal to noise ratios (SNRs), with the SNR provided to the system at test time. 5 The total number of test mixtures for each subject was thus 195. 3 see http://amp.ece.cmu.edu/projects/ Audio VisualSpeechProcessing/ 4Sine windows were used in analysis and synthesis such that their product forms windows that sum to unity when overlapped 50%. The windowed frames were analyzed using a 264-point fast Fourier transform (FFT) . The phases of the resulting spectra were discarded. 5Estimation of the SNR is necessary in practice; however this subject has been treated The separated test sounds were estimated by the system under two conditions: with and without the use of video information. We evaluated the estimates on the test set using a speech recognition system developed by Bhiksha Raj, using the eMU Sphinx ASR engine. 6 Existing speech HMMs trained on 60 hours of broadcast news data were used for recognition. 7 The models were adapted in an unsupervised manner to clean speech from each speaker, by learning a single affine transformation of all the state means, using a maximum likelihood linear regression procedure [14]. The recognizer adapted to each speaker was tested with the enhanced speech produced by the speech model for that speaker, as well as with no enhancement. Results are shown in figure 4(b). Recognition was greatly facilitated by the enhancement, with additional gains resulting from the use of video. It is somewhat surprising that the gains for video occur mostly in areas of higher SNR, whereas in human speech perception they occur under lower SNR. Little subjective difference was noted with the use of video in the case of two speakers. However in other experiments, when both voices came from the same speaker, the video was crucial in disambiguating which signal came from which voice. "one" "two" Originals Mixture Separated SNR dB (a) signal separation spectrograms (b) automatic speech recognition Figure 4: spectrograms of separated speech signals for a mixture two words spoken by the same speaker (a), and speech recognition performance for 39 mixtures of two words spoken by different speakers (b) 6 Discussion We have presented promising techniques for audio-visual speech enhancement. We introduced a factorial HMM to track both formant and pitch information, as well as video, in a unified probabilistic model, and demonstrated its effectiveness in speech enhancement. We are not aware of any other HMM-based audio-visual elsewhere [6] and is beyond the scope of this paper. 6 see http://www.speech.cs.cmu.edu/sphinxj. 7These models represented every combination of three phones (triphones) using 6000 states tied across trip hone models, with a 16-element Gaussian mixture observation model for each state. The data were processed at 8 kHz in 25ms windows overlapped by 15ms, with a frame rate of 100 frames per second, and analyzed into 31 Mel frequency components from which 13 cepstral coefficients were derived. These coefficients with the mean vector removed, and supplemented with their time differences, comprised the observed features speech enhancement systems in the literature. The results are tentative given the small sample of voices used ; however they suggest that further study with a larger sample of voices is warranted. It would be useful to compare the performance of a factorial speech model to that of each factor in isolation, as well as to a fullspectrum model. Measures of quality and intelligibility by human listeners in terms of speech and emotion recognition , as well as speaker identity, will also be helpful in further demonstrating the utility of these techniques. We look forward to further development of these techniques in future research. Acknowledgments We wish to thank Mitsubishi Electric Research Labs for hosting this research. Special thanks to Bhiksha Raj for devising and producing the evaluation using speech recognition , and to Matt Brand for his entropic HMM toolkit. References [1] W. H. Sumby and I. Pollack. Visual contribution to speech intelligibility in noise. Journal of th e Acoustical Society of America, 26:212- 215, 1954. [2] Jordi Robert-Ribes, Jean-Luc Schwartz, Tahar Lallouache, and Pierre Escudier. Complementarity and synergy in bimodal speech. Journ el of the Acoustical Society of America, 103(6):3677- 3689, 1998. [3] Stepmane Dupont and Juergen Luettin. Audio-visual speech modeling for continuous speech recognition. IEEE transactions on Multimedia, 2(3):141- 151, 2000. [4] John W. Fisher, Trevor Darrell , William T. Freeman, and Paul Viola. Learning joint statistical models for audio-visual fusion and segregation. In Advances in Neural Information Processing Systems 13. 200l. [5] Laurent Girin , Jean-Luc Schwartz, and Gang Feng. Audio-visual enhancement of speech in noise. Journ el of the Acoustical Society of America, 109(6):3007- 3019, 200l. [6] Yariv Ephraim. Statistical-model based speech enhancement systems. Proceedings of th e IEEE, 80(10):1526- 1554, 1992. [7] Z. Ghahramani and M. Jordan. Factorial hidden markov models. In David S. Touretzky, Michael C. Mozer , and M.E. Hasselmo, editors, Advances in Neural Information Processing Systems 8, 1996. [8] Sam T. Roweis. One microphone source separation. In Advances in Neural Information Processing Systems 13. 200l. [9] Hagai Attias, John C. Platt , Alex Acero, and Li Deng. Speech denoising and dereverb eration using probabilistic models. In Advances in Neural Information Processing Systems 13. 200l. [10] M. J. F . Gales. Mod el-Bas ed Techniques for Noise Robust Speech R ecognition. PhD thesis, Cambridge University, 1996. [11] F . R. Kschischang, B. Frey, and H .-A. Loeliger. Factor graphs and the sum-product algorithm. IEEE Trans. Inform. Theory, 47(2):498- 519, 200l. [12] F. J. Huang and T. Chen. Real-time lip-synch face animation driven by human voice. In IEEE Wo rkshop on Multimedia Signal Processing, Los Angeles, California, Dec 1998. [13] Matt Brand. Structure learning in conditional probability models via an entropic prior and parameter extinction. Neural Computation, 11(5):1155- 1182, 1999. [14] C. J. Leggetter and P. C. Woodland. Maximum likelihood linear regression for speaker adaptation of the parameters of continuous density hidden markov models. Computer Speech and Language, 9: 171- 185, 1995.
2005 |@word middle:1 manageable:1 extinction:2 open:1 mitsubishi:2 covariance:4 carry:1 series:1 contains:1 loeliger:1 envision:1 amp:1 subjective:1 existing:1 current:3 com:1 anne:1 surprising:1 must:1 john:3 subsequent:1 subcomponent:1 dupont:1 discrimination:1 stationary:2 cue:7 half:1 selected:1 devising:1 plane:1 provides:2 monday:1 five:1 height:2 mathematical:1 along:1 windowed:1 become:1 combine:2 manner:2 introduce:1 expected:3 multi:1 formants:2 audiovisual:1 freeman:1 little:2 window:3 provided:1 estimating:1 project:1 what:1 pursue:1 developed:3 spoken:3 unified:1 transformation:1 every:2 schwartz:2 platt:1 appear:2 producing:1 frey:1 laurent:1 plus:1 twice:1 suggests:2 hmms:12 ease:1 wideband:1 range:1 acknowledgment:1 yariv:1 practice:3 procedure:4 area:2 composite:10 word:12 distracter:1 suggest:1 clever:1 acero:1 context:1 www:1 outweigh:1 map:1 demonstrated:2 baum:1 uttered:1 independently:1 welch:1 simplicity:1 estimator:1 rule:1 his:1 variation:1 diego:1 enhanced:3 play:2 target:3 trick:1 element:2 overlapped:2 recognition:13 expensive:1 approximated:1 complementarity:1 database:2 bottom:2 role:1 recog:1 observed:1 calculate:1 news:1 removed:1 ephraim:2 mozer:1 predictable:1 complexity:3 dynamic:6 trained:4 basis:1 binaural:1 easily:1 joint:1 slate:1 represented:4 listener:1 america:3 train:2 separated:5 snrs:1 fast:1 cogsci:2 tell:1 whose:3 modular:1 larger:1 jean:2 formant:7 transform:1 noisy:3 itself:1 sequence:5 advantage:1 propose:1 product:3 adaptation:1 neighboring:1 combining:3 mixing:1 roweis:1 los:1 enhancement:16 regularity:1 darrell:1 produce:1 converges:1 derive:1 develop:1 auxiliary:1 c:1 quantify:1 differ:1 closely:1 filter:5 human:4 enable:1 bin:1 preliminary:2 hagai:1 tracker:1 normal:3 equilibrium:1 scope:1 vary:1 early:1 entropic:3 resample:1 recognizer:2 estimation:3 combinatorial:1 vibration:1 hasselmo:1 clearly:2 gaussian:5 rather:2 avoid:1 sparsify:1 publication:1 derived:4 focus:2 casey:2 improvement:2 likelihood:4 mainly:1 greatly:1 contrast:1 tremendously:1 glass:1 helpful:2 inference:6 dependent:2 el:3 typically:4 integrated:1 hidden:4 journ:2 among:1 dual:1 development:2 spatial:1 integration:2 initialize:1 mutual:1 special:1 emotion:2 construct:2 asr:4 comprise:1 having:1 aware:1 look:1 unsupervised:1 foreground:1 future:1 serious:1 employ:2 opening:1 randomly:1 resulted:1 parsimoniously:1 phase:8 vowel:2 william:1 evaluation:2 male:1 mixture:15 analyzed:3 yielding:2 amenable:1 accurate:1 ambient:1 fu:1 explosion:1 necessary:2 filled:1 loosely:3 re:2 desired:1 obscured:2 isolated:1 pollack:1 merl:1 lwe:1 modeling:3 instance:2 juergen:1 cost:1 snr:5 comprised:1 wonder:1 dependency:1 combined:5 person:1 density:4 thanks:1 probabilistic:2 michael:2 enhance:2 quickly:1 together:1 synthesis:1 thesis:1 ear:1 containing:1 huang:2 broadcast:1 gale:1 cognitive:1 corner:3 sphinx:1 derivative:1 li:1 summarized:1 includes:1 coefficient:2 sine:1 lab:2 closed:1 defer:1 contribution:1 wiener:1 variance:11 largely:2 phoneme:1 characteristic:1 yield:1 ofthe:1 fhmm:4 iterated:1 produced:2 marginally:1 simultaneous:1 inform:1 touretzky:1 trevor:1 ed:1 frequency:5 conveys:1 associated:1 di:1 jordi:1 gain:2 auditory:1 revise:1 conversation:1 knowledge:2 amplitude:4 reflecting:2 higher:1 follow:1 hershey:1 fhmms:2 improved:1 specify:1 formulation:1 evaluated:1 until:1 ally:1 night:1 overlapping:2 propagation:1 bly:1 quality:2 escudier:1 stale:1 bhiksha:2 matt:2 consisted:1 hence:1 read:1 conditionally:1 numerator:1 width:2 during:3 speaker:17 noted:1 mel:1 m:2 impression:1 delivers:1 variational:2 novel:1 recently:1 khz:3 million:1 extend:1 mellon:1 cambridge:2 framed:1 automatic:4 similarly:1 language:1 had:1 toolkit:1 longer:1 etc:1 add:3 subjectively:1 triphones:1 female:1 raj:2 moderate:1 irrelevant:1 phone:1 scenario:1 phonetic:1 driven:1 came:2 nition:1 seen:1 additional:2 somewhat:1 spectrogram:8 employed:1 deng:1 converge:1 maximize:1 signal:30 full:4 sound:8 infer:1 smooth:1 match:3 adapt:1 offer:1 pitch:11 regression:2 denominator:1 vision:5 enhancing:1 cmu:2 iteration:1 represent:2 bimodal:1 luettin:1 dec:1 background:4 remarkably:1 addition:1 whereas:2 separately:1 source:2 crucial:2 recording:2 tend:2 subject:5 db:2 mod:1 effectiveness:1 jordan:1 near:1 enough:1 fft:1 variety:1 isolation:1 architecture:1 reduce:2 prototype:1 attias:1 angeles:1 pca:1 utility:2 granted:1 effort:1 wo:1 suffer:1 speech:74 cocktail:1 jie:1 useful:3 woodland:1 detailed:1 factorial:14 hosting:1 transforms:2 nonparametric:1 revived:1 amount:1 band:35 processed:1 simplest:1 http:2 estimated:3 per:3 track:1 carnegie:1 demonstrating:1 k4:2 neither:1 clean:5 backward:5 graph:1 sum:11 facilitated:2 uncertainty:2 saying:1 submodels:1 separation:6 bit:1 entirely:1 layer:1 yielded:1 adapted:2 occur:2 gang:1 handful:1 constrain:1 alex:1 cepstra:2 sake:1 interpolated:1 fourier:1 extremely:1 department:1 structured:1 according:1 combination:6 beneficial:1 across:3 slightly:1 sam:1 unity:1 explained:1 segregation:1 mechanism:1 apply:1 intelligibility:3 spectral:6 pierre:1 voice:10 original:2 top:1 remaining:1 graphical:1 music:1 exploit:1 ghahramani:1 february:1 society:3 contact:1 feng:1 added:2 question:1 strategy:2 diagonal:3 ecognition:1 separate:1 thank:1 separating:3 hmm:18 outer:1 chris:1 acoustical:4 reason:1 length:1 ratio:1 guest:1 difficult:1 mostly:1 robert:1 holding:1 reverberation:1 ba:1 unknown:1 upper:3 observation:29 markov:4 discarded:3 hone:1 viola:1 incorporated:2 frame:15 ucsd:2 inferred:1 introduced:1 david:1 pair:2 required:1 trip:1 coalesce:1 tentative:1 california:2 acoustic:2 learned:3 narrow:19 engine:1 emu:1 hour:1 trans:1 beyond:1 below:1 perception:2 pattern:2 articulation:1 hear:1 challenge:1 including:2 video:20 belief:1 mouth:1 overlap:2 eration:1 treated:1 recursion:2 clink:1 scheme:1 jhershey:2 utterance:2 coupled:2 prior:5 literature:1 relative:2 mixed:1 interesting:1 filtering:2 utter:1 affine:1 sufficient:1 imposes:1 editor:1 course:1 elsewhere:1 allow:1 perceptron:1 wide:20 face:3 cepstral:4 lognormal:3 factorially:3 taking:2 unmanageable:1 benefit:3 distributed:1 dimension:1 vocabulary:2 transition:2 sumby:1 contour:1 forward:6 commonly:2 made:1 san:1 party:1 employing:1 transaction:1 approximate:3 ignore:1 synergy:1 sequentially:1 consonant:1 spectrum:10 continuous:2 lip:10 promising:2 robust:1 kschischang:1 obtaining:1 warranted:1 complex:3 electric:2 domain:6 synch:1 noise:19 massachussets:1 paul:1 animation:1 repeated:1 complementary:1 depicts:4 sub:10 wish:1 tied:1 late:1 removing:1 formula:1 showing:1 supplemented:1 disregarding:1 fusion:1 burden:1 incorporating:1 naively:1 consist:1 adding:2 phd:1 illustrates:1 chen:1 easier:1 simply:1 likely:3 visual:25 extracted:2 conditional:1 modulate:1 identity:2 formulated:1 room:1 disambiguating:1 luc:2 fisher:1 included:1 uniformly:2 denoising:1 microphone:2 total:1 multimedia:2 pas:2 ece:1 brand:2 searched:1 prosody:1 incorporate:2 evaluate:2 audio:28 tested:1
1,103
2,006
. Information-geometric decomposition In spike analysis Hiroyuki Nakahara; Shun-ichi Amari Lab. for Mathematical Neuroscience, RIKEN Brain Science Institute 2-1 Hirosawa, Wako, Saitama, 351-0198 Japan {him, amari} @brain.riken.go.jp Abstract We present an information-geometric measure to systematically investigate neuronal firing patterns, taking account not only of the second-order but also of higher-order interactions. We begin with the case of two neurons for illustration and show how to test whether or not any pairwise correlation in one period is significantly different from that in the other period. In order to test such a hypothesis of different firing rates, the correlation term needs to be singled out 'orthogonally' to the firing rates, where the null hypothesis might not be of independent firing. This method is also shown to directly associate neural firing with behavior via their mutual information, which is decomposed into two types of information, conveyed by mean firing rate and coincident firing, respectively. Then, we show that these results, using the 'orthogonal' decomposition, are naturally extended to the case of three neurons and n neurons in general. 1 Introduction Based on the theory of hierarchical structure and related invariant decomposition of interactions by information geometry [3], the present paper briefly summarizes methods useful for systematically analyzing a population of neural firing [9]. Many researches have shown that the mean firing rate of a single neuron may carry significant information on sensory and motion signals. Information conveyed by populational firing, however, may not be only an accumulation of mean firing rates. Other statistical structure, e.g., coincident firing [13, 14], may also carry behavioral information. One obvious step to investigate this issue is to single out a contribution by coincident firing between two neurons, i.e., the pairwise correlation [2, 6]. In general, however, it is not sufficient to test a pairwise correlation of neural firing, because there can be triplewise and higher correlations. For example, three variables (neurons) are not independent in general even when they are pairwise independent. We need to establish a systematic method of analysis, including these higher-order ? also affiliated with Dept. of Knowledge Sci., Japan Advanced Inst. of Sci. & Tech. correlations [1, 5,7, 13] . We propose one approach, the information-geometric measure that uses the dual orthogonality of the natural and expectation parameters in exponential family distributions [4]. We represent a neural firing pattern by a binary random vector x. The probability distribution of firing patterns can be expanded by a log linear model, where the set {p( x)} of all the probability distributions forms a (2n - I)-dimensional manifold 8 n . Each p(x) is given by 2n probabilities pi1???in=Prob{X1=i1,???,Xn=in}, ik=O,I, subjectto L Pi1???in=1 il ,"',i n and expansion in log p( x) is given by logp(x) = L BiXi + L BijXiXj + L i<j BijkXiXjXk??? + B1... nX1 ... Xn - 'Ij;, i<j<k where indices of Bijk, etc. satisfy i < j < k, etc. We can have a general theory of this n neuron case [3, 9], however , to be concrete given the limited space, we mainly discuss two and three neuron cases in the present paper. Our method shares some features with previous studies (e.g. [7]) in use of the log linear model. Yet, we make explicit use of the dual orthogonality so that the method becomes more transparent and more systematic. In the present paper, we are interested in addressing two issues: (1) to analyze correlated firing of neurons and (2) to connect such a technique with behavioral events. In (1), previous studies often assumed independent firing as the null hypothesis. However, for example, when we compare firing patterns in two periods, as control and 'test' periods, there may exist a weak correlation in the control period. Hence, benefiting from the 'orthogonal' coordinates, we develop a method applicable to the null hypothesis of non-independent firing, irrespective of firing rates. It is equally important to relate such a method with investigation of behavioral significance as (2). We show that we can do so, using orthogonal decomposition of the mutual information (MI) between firing and behavior [11, 12]. In the following , we discuss first the case of two neurons and then the case of three neurons , demonstrating our method with artificial simulated data. The validity of our method has been shown also with experimental data[9, 10] but not shown here due to the limited space. 2 Information-geometric measure: case of two neurons ? We denote two neurons by Xl and X 2 (Xi = 1, indicates if neuron i has a spike or not in a short time bin). Its joint probability p(x), x = (X1,X2), is given by Pij = Prob{x1 = i;X2 = j} > 0, i,j = 0, 1. Among four probabilities, {POO ,P01,P10,Pl1}, only three are free. The set of all such distributions of x forms a three-dimensional manifold 8 2. Any three of Pij can be used as a coordinate system of 8 2. There are many different coordinate systems of 8 2 . The coordinates of the expectation parameters, called 17-coordinates, 'TI = (171,172,1712), is given by 17i = Prob {Xi = I} = E [Xi], i = 1,2, 173 = 1712 = E [X1 X2] = P12, where E denotes the expectation and 17i and 1712 correspond to the mean firing rates and the mean coincident firing, respectively. As other coordinate systems, we can also use the triplet, (171,172, Cov [Xl, X 2]) , where Cov [Xl , X 2] is the covariance,and/or the triplet (171,172, p), where p is the correlation coefficient (COR), p = J '112 -,/11 '12 , often called N-JPSTH [2]. '/11 (l - '7d'72 (1 - '72) Which quantity would be convenient to represent the pairwise correlational component? It is desirable to define the degree of the correlation independently from the marginals (171,172), To this end, we use the 'orthogonal' coordinates (171 , 172 , B), originating from information geometry of 8 2 , so that the coordinate curve of B is always orthogonal to those of 171 and 172. The orthogonality of two directions in 8 2 (8 n in general) is defined by the Riemannian metric due to the Fisher information matrix [8, 4]. Denoting any coordinates in 8 n by ~ = (6, ... , ~n)' the Fisher information matrix G is given by (1) where l 9ij(~) (x;~) = logp (x; ~). The orthogonality between ~i and = O. In case of 8 2 , we desire to have E [tel (X ;171 , 172, B) ~j is defined by 8~il(x;171'172,B)] = o (i = 1, 2). When B is orthogonal to (171, 172), we say that B represents pure correlations independently of marginals. Such B is given by the following theorem. Theorem 1. The coordinate B = log PuPoo P01PlO is orthogonal to the marginals 171 and 172 . (2) We have another interpretation of B. Let's expand p(x) by logp(x) = L;=l BiXi + B12X1X2 - 'IjJ. Simple calculation lets us get the coefficients, B1 = log Pia, B2 = paa log EQl, ' I jJ = -logpoo, and B = B12 (as Eq 2). The triplet () = (B1' B2, B ) forms 12 paa another coordinate system, called the natural parameters, or B-coordinates. We remark that B12 is 0 when and only when Xl and X 2 are independent. The triplet C== (171,172,B12 ) forms an 'orthogonal' coordinate system of 8 2 , called the mixed coordinates [4]. We use the Kullback-Leibler divergence (KL) to measure the discrepancy between two probabilities p(x) and q(x) , defined by D[p:q] = LxP(x)log~t~}. In the following , we denote any coordinates of p by etc (the same for q). Using the orthogonality between 17- and B-coordinates, we have the decomposition in the KL. e Theorem 2. D [q : p] = D [q : r**] + D [r** : p] , + D [r* : q], (3) are given by Cr = (17f, 17~, Bj) and Cr = (17f, 17g, B~), respectively. D [p : q] = D [p : r*] where r* and r** > ? The squared distance ds 2 between two nearby distributions p(x , ~) and p(x,~, +d~) is given by the quadratic form of d~, L ds 2 = 9ij(~)d~id~j, i,jE(1,2,3) which is approximately twice the KL, i.e. , ds 2 ~ 2D [P(x , ~) : p(x,~ Now suppose ~ is the mixed coordinates is of the form gfj =[ gll g~2 gl2 gf2 o 0 0 0 g~3 C. + ~)]. Then, the Fisher information matrix 1and we have ds 2 = dsi g~3(dB3) 2, ds~ = Li,j E(1,2) 9fjd17id17j, corresponding to Eq. 3. + ds~, where dsi = This decomposition comes from the choice of the orthogonal coordinates and gives us the merits of simple procedure in statistical inference. First, let us estimate the parameter TI = (1}1,1}2) and B from N observed data Xl, ... , XN. The maximum likelihood estimator (mle) ( , which is asymptotically unbiased and efficient, is easily obtained by 1)' . = l..#{x? = I} and 8 = log fh?(1-=-fh-.ib+~12) using ? N (1]1-1]12)(1]2-1]12) , ? fj12 = tt#{XIX2 = I}. The covariance of estimation error, f::J.TI and f::J.B, is given asymptotically by Cov [ ~~ ] = ttGZ1. Since the cross terms of G or G- 1 vanish for the orthogonal coordinates, we have Cov [f::J.TI, f::J.B] = 0, implying that the estimation error f::J.TI of marginals and that of interaction are mutually independent. Such a property does not hold for other non-orthogonal parameterization such as the COR p, the covariance etc. Second, in practice, we often like to compare many spike distributions, q(x(t)) (i.e, (q(t)) for (t = 1", T), with a distribution in the control period p( x) , or (P. Because the orthogonality between TI and B allows us to treat them independently, these comparisons become very simple. These properties bring a simple procedure of testing hypothesis concerning the null hypothesis against (4) Ho : B = Bo where Bo is not necessarily zero, whereas Bo = 0 corresponds to the null hypothesis of independent firing , which is often used in literature in different setting. Let the log likelihood of the models Ho and HI be, respectively, lo = maxlogp(Xl ' ... , XN ; TI , Bo) TI and h = maxlogp(Xl' ... , XN; TI, B). TI,e The likelihood ratio test uses the test statistics A = 2log ~. By the mle with respect to TI and which can be performed independently, we have e, (5) lo = logp(x ,r"B o), where r, are the same in both models. A similar situation holds in the case of testing TI = Tlo against TI =I Tlo for unknown B. Under the hypothesis H o, A is approximated for a large N as A= 2 t log i=l P(Xi;~' B~) ';::;j N gi3 (8 - BO)2 '" X2(1). (6) p(Xi; TI, B) Thus, we can easily submit our data to a hypothetical testing of significant coincident firing against null hypothesis of any correlated firing, independently from the mean firing rate modulation 1 . We now turn to relate the above approach with another important issue, which is to relate such a coincident firing with behavior. Let us denote by Y a variable of discrete behavioral choices. The MI between X = (X1 ,X2 ) and Y is written by J(X, Y) = Ep(x ,y) p(x , y)] [ log p(x)p(y) = Ep(Y) [D [P(Xly) : p(X)]]. Using the mixed coordinates for p(Xly) and p(X) , we have D [P(Xly) : p(X)] D [?(Xly) : ?(X)] = D [?(Xly) : ('J + D [(I : ?(X)J, where (' = ('(X,y) ((1 (Xly), (2 (X Iy) , (3 (X)) = (1}1 (Xly), 1}2(Xly), B3(X)). 1 A more proper formulation in this hypothetical testing can be derived, resulting in using p value from X2 (2) distribution , but we omit it here due to the limited space [9] Theorem 3. + h(X, Y) , J(X, Y) = It (X , Y) where It (X , Y), h(X, Y) are given by It (X, Y) = Ep(Y) [D [?(Xly) : ('(X,y)]] ,h(X, Y) = (7) Ep(Y) [D [('(X,y) : ?(X)]] . Obviously, the similar result holds with respect to p(YIX). By this theorem, J is the sum of the two terms: It is by modulation of the correlation components of X, while h is by modulation of the marginals of X. This observation helps us investigate the behavioral significance by modulating either coincident firing or mean firing rates. 0.1 0 . 1 ,-----~-~--~-~-____, ~ A (al / 'J, ~ 0.05~ .?.??.? __ I~)? ? _ ~, .? .? .?.? .?.? .? ~C)? ? .? .? .? .? .? u; o a 00 - ........................ _12 . .. - 100 300 500 100 300 lime Ims) 500 ,-------~--~-~-____, B \ll . , .. , .. . , . .. , .. ". , .. ' .. , .. " . . .. ' .. , . ... . . , .. . . , .. , _ .# ' ...... .. , .. . . ' .. -" .. . , ?? , !- ~0.05 W : 112 ~ _________ 'J , ~ .. __ '_12 '- - - .-- - --- -"" -_ .. ?0:------=:-: 10:::0 ~~~~3=00=---~---:-: 500 00 100 "300 lime Ims) 500 Figure 1: Demonstration of information-geometric measure in two neuron case, using simulated neural data, where two behavioral choices (sl, s2) are assumed. A,B. (1]1 , 1]2 , 1]12) with respect to sl, s2. C,D . COR,B, computed by using ", L-iP(Si)",(Si) with P(Si) = 1/2 (i = 1, 2). E. p-values. F. MI. Fig 1 succinctly demonstrates results in this section. Figs 1 A, B are supposed to show mean firing rates of two neurons and mean coincident firing for two different stimuli (sl, s2). The period (a) is assumed as the control period, i.e. , where no stimuli is shown yet, whereas the stimulus is shown in the periods (b,c). Fig 1 C, D gives COR, B. They look to change similarly over periods, which is reasonable because both COR and B represent the same correlational component, but indeed change slightly differently over periods (e.g., the relative magnitudes between the periods (a) and (c) are different for COR and B) , which is also reasonable because both represent the correlational component as in different coordinate systems. Using B in Fig 1 D, Fig 1 E shows p-values derived from X2 (1) (i.e., P > 0.95 in Fig 1 E is 'a significance with P < 0.05') for two different null hypotheses , one of the averaged firing in the control period (by solid line) and the other of independent firing (by dashed line) , which is of popular use in literature. In general, it becomes complicated to test the former hypothesis , using COR. This is because the COR, as the coordinate component, is not orthogonal to the mean firing rates so that estimation errors among the COR and mean firing rates are entangled and that the proper metric among them is rather difficult to compute. Once using B, this testing becomes simple due to orthogonality between B and mean firing rates. Notably, we would draw completely different conclusions on significant coincident firing given each null hypothesis in Fig 1 E. This difference may be striking when we are to understand the brain function with these kinds of data. Fig 1 F shows the MI between firing and behavior, where behavioral event is with respect to stimuli, and its decomposition. There is no behavioral information conveyed by the modulation of coincident firing in the period (b) (i.e., h = 0 in the period (b)). The increase in the total MI (i.e., I) in the period (c), compared with the period (b), is due not to the MI in mean firing (h) but to the MI correlation (h). Thus, with a great ease, we can directly inspect a function of neural correlation component in relation to behavior. 3 Three neuron case With more than two neurons, we need to look not only into a pairwise interaction but also into higher-order interactions. Our results in the two neuron case are naturally extended to n neuron case and here, we focus on three neuron case for illustration. For three neurons X = (X 1,X2,X3), we let p(x), x = (X1,X2,X3), be their joint probability distribution and put Pijk = Prob {Xl = i, X2 = j, X3 = k}, i, j, k = 0,1. The set of all such distributions forms a 7-dimensional manifold 8 3 due to "L.Pijk = 1. The 1]-coordinates 'fI = ('fI1; 'fI2; 'fI3) = (1]1,1]2,1]3; 1]12,1]23,1]13; 1]123) is defined by (i, j = 1,2, 3; i i- j), 1]123 = E [X1X2X3]. To single out the purely triplewise correlation, we utilize the dual orthogonality of 8- and 1]-coordinates. By using expansion of log p( x) = "L. 8iXi + "L.8ijXiXj + 8123X1X2X3 - 'ljJ, we obtain 8-coordinates, () = (()1;()2;()3) = (8 1,82,83; 812 ,823 ,8 13 ; 8123 ). It's easy to get the expression of these coefficients (e.g. ,123 8 = log P110PIOIP0l1POOO P111 PIOO POIOP001). Information geometry gives the following theorem. 1]i = E [Xi] (i = 1,2,3), 1Jij = E [XiXj] Theorem 4. 8123 represents the pure triplewise interaction in the sense that it is orthogonal to any changes in the single and pairwise marginals, i.e., 'fIl and 'fI2. We use the following two mixed coordinates to utilize the dual orthogonality, (I = ('fIl; ()2; ()3), (2 = ('fIl; 'fI2; ()3). Here (2 is useful to single out the triple wise interaction (()3 = 8123 ), while (I is to single out the pairwise and triplewise interactions together (()2; ()3). Note that 8123 is not orthogonal to {8 ij }. In other words , except the case of no triple wise interaction (8 123 = 0), 8ij do not directly represent the pairwise correlation of two random variables Xi, X j . The case of independent firing is given by 1]ij = 1]i1]j, 1]123 = 1]11]21]3 or equivalently by ()2 = 0, ()3 = o. The decomposition in the KL is now given as follows. Theorem 5. D [p : q] = D [p : p] + D [p : q] = D [p : fi] + D [p : q] = D [p : p] + D [p : fi] + D [p : q] . where, using the mixed coordinates, we have (g = (8) ('fIi; 'fI~; ()?), (f = ('fIi; ()~; ()?). A hypothetical testing is formulated similarly to the two neuron case. We can examine a significance of the triplewise interaction by A2 = 2ND [p : p] ~ N g~7 (~) (8f238i23)2 ~ X2(1). For a significance of triplewise and pairwise interactions together, we have Al = 2ND [p : fi] ~ N "L.J,j=4 gfj(f)((f - (f)((f - (f) ~ X2(4). For the decomposition of the MI between firing X and behavior Y, we have Theorem 6. J(X, Y) = h (X, Y) + h(X, Y) = h(X, Y) + J4(X, Y) (9) where h(X, Y) = Ep(Y) [D [( I(X ly ) : ( I(X,y)] ] , h(X, Y) = Ep(Y) [D [(I (X,y) : ( I(X) ]] , h(X, Y) = Ep(Y) [D [(2(X ly ) : ( 2(X,y)] ] , Ep(Y) [D [( 2(X,y) : ( 2(X) ]], 14 (X, Y) = By t he first equality, I is decomposed into two parts: II is conveyed by the pairwise and triplewise interactions of firing, and h by the mean firing rate modulation. By the second equality, I is decomposed differently: h, conveyed by t he triplewise interaction, and 14 , by the other terms. 00 (e) (a) 100 r:- ~ : - 8~~I----' [: 0.04 '"0.02 (d) 300 500 700 -0. 10 300 500 700 0 100 300 500 700 .. ~ ~~j ~.~~ it~,-~ I? ? -2 0 100 1- _____ _ N205 E x - - - i: -. -. --, --) ---- ." ,t: t' .....~I/"'\:""''! ~I ~....,:..;..".u,,'rV! '/ J.~'" ~ t. I oo'-'--"1"" oo:"-'-'-'-'-"'-'----::3~ 00c--~---'c5-:-:00-~---=1l700 time (ms) 95 100 300 500 700 N~05011-F- - - - - - - - - - - .':,f?- - - - ? - - - - - - - - .- - - - 95 ...... 0 100 300 time (ms) 500 700 Figure 2: Demonstration in three neuron case. A '11 = ('111> '112, '113) ~ ('T/i , 'T/ij,'T/ijk) from top to bottom, since we treated a homogeneous case in this simulation for simplicity. B. COR. C. (}12,(}13 , (}23' D (}12 3 . E p-value,...., X2 (1). F p-value,...., X2(4). We emphasize that all the above decompositions come from the choice of the 'ort hogonal' coordinates. Fig 2 highlights some of the results in this section. Fig 2 A shows t he mean firing rates (see legend). The period (a) is assumed as t he control period. Fig 2 B indicates t hat COR changes only in the periods (c,d), while Fig 2 C indicates that (}123 changes only in t he period (d). Taken together, we observe that t he triplewise correlation (}123 can be modulated independently from COR. Fig 2 E indicates the p-value from X2(1) against the null hypothesis of the activity in t he control period. The triplewise coincident firing becomes significant only in the period (d). Fig 2 F indicates the p-value from X2(4) . The coincident firing, taking t he triplewise and pairwise interaction together, becomes significant in both periods (c,d). We cannot observe these differences in modulation of pairwise and triplewise interactions over periods (c, d), when we inspect only COR. Remark: For a general n neuron case, we can use the k-cut mixed coordinates, (k = ('111 ' ... , '11 k; 0 k+l, .. . , On) = ('I1k- ; 0k+)' Using the orthogonality between 'I1kand 0 k+, the similar results hold. To meet the computational complexity involved in this general case, some practical difficulties should be resolved in practice [9] . 4 Discussions We presented the information-geometric measures to analyze spike firing patterns, using two and three neuron cases for illustration. The choice of 'orthogonal' coordinates provides us with a simple, transparent and systematic procedure to test significant firing patterns and to directly relate such a pattern with behavior. We hope that this method simplifies and strengthens experimental data analysis. Acknowledgments HN thanks M. Tatsuno, K. Siu and K. Kobayashi for their assistance. HN is supported by Grants-in-Aid 13210154 from the Ministry of Edu. Japan. References [1] M. Abeles, H. Bergman, E. Margalit, and E. Vaadia. Spatiotemporal firing patterns in the frontal cortex of behaving monkeys. J Neurophysiol, 70(4):162938.,1993. [2] A. M. H. J. Aertsen, G. 1. Gerstein, M. K. Habib, and G. Palm. Dynamics of neuronal firing correlation: Modulation of "effective connectivity". Journal of Neurophysiology, 61(5):900- 917, May 1989. [3] S. Amari. Information geometry on hierarchical decomposition of stochastic interactions. IEEE Transaction on Information Theory, pages 1701- 1711,2001. [4] S. Amari and H. Nagaoka. Methods of Information Geometry. AMS and Oxford University Press, 2000. [5] S. Griin. Unitary joint-events in multiple-neuron spiking activity: detection, significance, and interpretation. Verlag Harri Deutsch, Reihe Physik, Band 60 . Thun, Frankfurt/Main, 1996. [6] H. Ito and S. Tsuji. Model dependence in quantification of spike interdependence by joint peri-stimulus time histogram. Neural Computation, 12:195- 217, 2000. [7] L. Martignon, G. Deco, K. Laskey, M. Diamond, W. A. Freiwald, and E. Vaadia. Neural coding: Higher-order temporal patterns in the neurostatistics of cell assemblies. Neural Computation, 12(11):2621- 2653, 2000. [8] H. Nagaoka and S. Amari. Differential geometry of smooth families of probability distributions. Technical report , University of Tokyo, 1982. [9] H. Nakahara and S. Amari. Information geometric measure for neural spikes. in prepration. [10] H. Nakahara, S. Amari, M. Tatsuno, S. Kang, K. Kobayashi , K. Anderson, E. Miller, and T. Poggio. Information geometric measures for spike firing. Society for Neuroscience Abstracts, 27:821.46 (page.2178), 2001. [11] M. W . Oram, N. G. Hatsopoulos, B. J. Richmond, and J . P. Donoghue. Excess synchrony in motor cortical neurons provides redundant direction information with that from coarse temporal measures. J Neurophysiol., 86(4):1700- 1716, 2001. [12] S. Panzeri and S. R. Schultz. A unified approach to the study of temporal, correlational, and rate coding. Neural Computation, 13(6):1311-49., 2001a. [13] A. Riehle, S. Griin, M. Diesmann, and A. Aertsen. Spike synchronization and rate modulation differentially involved in motor cortical function. Science, 278:1950- 1953, 12 Dec 1997. [14] E. Vaadia, I. Haalman, M. Abeles, H. Bergman, Y. Prut, H. Slovin, and A. Aertsen. Dynamics of neuronal interactions in monkey cortex in relation to behavioural events. Nature, 373:515- 518, 9 Feb 1995.
2006 |@word neurophysiology:1 briefly:1 nd:2 physik:1 simulation:1 decomposition:11 covariance:3 solid:1 carry:2 tlo:2 denoting:1 wako:1 si:3 yet:2 written:1 motor:2 implying:1 parameterization:1 short:1 coarse:1 provides:2 bixi:2 mathematical:1 become:1 differential:1 ik:1 behavioral:8 interdependence:1 pairwise:13 notably:1 indeed:1 behavior:7 examine:1 brain:3 decomposed:3 bijk:1 becomes:5 begin:1 null:9 kind:1 monkey:2 unified:1 temporal:3 hypothetical:3 ti:14 demonstrates:1 control:7 ly:2 omit:1 grant:1 kobayashi:2 treat:1 analyzing:1 id:1 oxford:1 meet:1 firing:54 modulation:8 approximately:1 might:1 twice:1 b12:3 gll:1 limited:3 ease:1 averaged:1 practical:1 acknowledgment:1 testing:6 practice:2 x3:3 procedure:3 significantly:1 convenient:1 word:1 get:2 cannot:1 put:1 accumulation:1 poo:1 go:1 independently:6 simplicity:1 pure:2 freiwald:1 estimator:1 population:1 coordinate:30 suppose:1 homogeneous:1 us:2 ixi:1 hypothesis:13 bergman:2 associate:1 approximated:1 strengthens:1 cut:1 observed:1 ep:8 bottom:1 db3:1 hatsopoulos:1 complexity:1 dynamic:2 purely:1 completely:1 neurophysiol:2 easily:2 joint:4 resolved:1 differently:2 harri:1 riken:2 effective:1 artificial:1 griin:2 fi3:1 say:1 amari:7 cov:4 statistic:1 nagaoka:2 singled:1 ip:1 obviously:1 vaadia:3 propose:1 interaction:17 jij:1 benefiting:1 supposed:1 differentially:1 help:1 oo:2 develop:1 ij:7 eq:2 come:2 xly:9 deutsch:1 direction:2 tokyo:1 stochastic:1 shun:1 bin:1 transparent:2 investigation:1 hold:4 fil:3 great:1 panzeri:1 bj:1 a2:1 fh:2 estimation:3 applicable:1 him:1 modulating:1 hope:1 always:1 yix:1 rather:1 cr:2 derived:2 focus:1 indicates:5 mainly:1 likelihood:3 tech:1 richmond:1 sense:1 am:1 inst:1 inference:1 paa:2 margalit:1 relation:2 originating:1 expand:1 i1:2 interested:1 issue:3 dual:4 among:3 mutual:2 once:1 represents:2 look:2 discrepancy:1 report:1 stimulus:5 divergence:1 geometry:6 detection:1 investigate:3 poggio:1 orthogonal:15 logp:4 pia:1 addressing:1 saitama:1 siu:1 connect:1 spatiotemporal:1 abele:2 thanks:1 peri:1 systematic:3 together:4 iy:1 concrete:1 hirosawa:1 connectivity:1 squared:1 deco:1 hn:2 li:1 japan:3 account:1 b2:2 coding:2 coefficient:3 satisfy:1 performed:1 lab:1 analyze:2 complicated:1 synchrony:1 contribution:1 il:2 prut:1 miller:1 correspond:1 weak:1 against:4 martignon:1 involved:2 obvious:1 naturally:2 mi:8 riemannian:1 oram:1 popular:1 knowledge:1 hiroyuki:1 higher:5 formulation:1 anderson:1 correlation:17 d:6 laskey:1 b3:1 validity:1 unbiased:1 former:1 hence:1 equality:2 leibler:1 riehle:1 haalman:1 ll:1 assistance:1 m:2 tt:1 motion:1 p12:1 bring:1 wise:2 fi:5 spiking:1 jp:1 interpretation:2 he:8 marginals:6 ims:2 significant:6 frankfurt:1 similarly:2 j4:1 cortex:2 behaving:1 etc:4 ort:1 feb:1 fii:2 pl1:1 verlag:1 binary:1 p10:1 ministry:1 period:25 redundant:1 signal:1 dashed:1 ii:1 multiple:1 desirable:1 rv:1 smooth:1 technical:1 calculation:1 cross:1 concerning:1 equally:1 mle:2 expectation:3 metric:2 histogram:1 represent:5 cell:1 dec:1 whereas:2 entangled:1 p01:1 legend:1 unitary:1 easy:1 simplifies:1 donoghue:1 whether:1 expression:1 jj:1 remark:2 useful:2 band:1 sl:3 exist:1 neuroscience:2 discrete:1 ichi:1 four:1 demonstrating:1 utilize:2 asymptotically:2 sum:1 prob:4 striking:1 family:2 reasonable:2 draw:1 gerstein:1 summarizes:1 lime:2 hi:1 quadratic:1 pijk:2 activity:2 orthogonality:10 i1k:1 x2:16 nearby:1 diesmann:1 pi1:2 expanded:1 palm:1 neurostatistics:1 slightly:1 xixj:1 invariant:1 taken:1 behavioural:1 mutually:1 discus:2 turn:1 merit:1 gfj:2 cor:13 end:1 fi1:1 observe:2 hierarchical:2 subjectto:1 ho:2 hat:1 denotes:1 top:1 assembly:1 establish:1 society:1 quantity:1 spike:8 dependence:1 aertsen:3 distance:1 sci:2 simulated:2 manifold:3 index:1 illustration:3 ratio:1 demonstration:2 equivalently:1 difficult:1 relate:4 affiliated:1 proper:2 unknown:1 diamond:1 inspect:2 neuron:28 observation:1 coincident:12 situation:1 extended:2 kl:4 kang:1 fi2:3 pattern:9 including:1 event:4 natural:2 treated:1 difficulty:1 quantification:1 advanced:1 orthogonally:1 irrespective:1 ljj:1 geometric:8 literature:2 relative:1 synchronization:1 dsi:2 highlight:1 mixed:6 triple:2 degree:1 conveyed:5 slovin:1 sufficient:1 pij:2 systematically:2 share:1 lo:2 succinctly:1 populational:1 supported:1 free:1 understand:1 institute:1 taking:2 curve:1 xn:5 cortical:2 sensory:1 c5:1 schultz:1 transaction:1 excess:1 emphasize:1 kullback:1 b1:3 assumed:4 xi:7 triplet:4 nature:1 tel:1 expansion:2 necessarily:1 submit:1 significance:6 main:1 s2:3 x1:6 neuronal:3 fig:14 je:1 aid:1 explicit:1 exponential:1 xl:8 ib:1 vanish:1 ito:1 theorem:9 hogonal:1 jpsth:1 magnitude:1 ijj:1 desire:1 bo:5 corresponds:1 lxp:1 xix2:1 formulated:1 nakahara:3 fisher:3 habib:1 change:5 except:1 correlational:4 called:4 total:1 experimental:2 ijk:1 nx1:1 modulated:1 frontal:1 dept:1 correlated:2
1,104
2,007
Quantizing Density Estimators Peter Meinicke Neuroinformatics Group University of Bielefeld Bielefeld, Germany pmeinick@techfak.uni-bielefeld.de Helge Ritter Neuroinformatics Group University of Bielefeld Bielefeld, Germany helge@techfak.uni-bielefeld.de Abstract We suggest a nonparametric framework for unsupervised learning of projection models in terms of density estimation on quantized sample spaces. The objective is not to optimally reconstruct the data but instead the quantizer is chosen to optimally reconstruct the density of the data. For the resulting quantizing density estimator (QDE) we present a general method for parameter estimation and model selection. We show how projection sets which correspond to traditional unsupervised methods like vector quantization or PCA appear in the new framework. For a principal component quantizer we present results on synthetic and realworld data, which show that the QDE can improve the generalization of the kernel density estimator although its estimate is based on significantly lower-dimensional projection indices of the data. 1 Introduction Unsupervised learning is essentially concerned with finding alternative representations for unlabeled data. These alternative representations usually reflect some important properties of the underlying distribution and usually they try to exploit some redundancy in the data. In that way many unsupervised methods aim at a complexity-reduced representation of the data, like the most common approaches, namely vector quantization (VQ) and principal component analysis (PCA). Both approaches can be viewed as specific kinds of quantization, which is a basic mechanism of complexity reduction. The objective of our approach to unsupervised learning is to achieve a suitable quantization of the data space which allows for an optimal reconstruction of the underlying density from a finite sample. In that way we consider unsupervised learning as density estimation on a quantized sample space and the resulting estimator will be referred to as quantizing density estimator (QDE). The construction of a QDE first requires to specify a suitable class of parametrized quantization functions and then to select from this set a certain function with good generalization properties. While the first point is common to unsupervised learning, the latter point is addressed in a density estimation framework where we tackle the model selection problem in a data-driven and nonparametric way. It is often overlooked that modern Bayesian approaches to unsupervised learning and model selection are almost always based on some strong assumptions about the data distribution. Unfortunately these assumptions usually cannot be inferred from human knowledge about the data domain and therefore the model building process is usually driven by computational considerations. Although our approach can be interpreted in terms of a generative model of the data, in contrast to most other generative models (see [10] for an overview), the present approach is nonparametric, since no specific assumptions about the functional form of the data distribution have to be made. In that way our approach compares well with other quantization methods, like principal curves and surfaces [4, 13, 6], which only have to make rather general assumptions about the underlying distribution. The QDE approach can utilize these methods as specific quantization techniques and shows a practical way how to further automatize the construction of unsupervised learning machines. 2 Quantization by Density Estimation We will now explain how the QDE may be derived from a generalization of the kernel density estimator (KDE), one of the most popular methods for nonparametric density estimation [12, 11]. If we construct a kernel density estimator on the basis of a quantized sample, we have the following estimator             (1)  $#%#& "! a sample from the target distribution, where   denotes the kernel ('*)+-is,/ .*0213)+  function and with parameter vector is a given quantization or  4  to a parametrized subset . 0 Sof the sample space projection function4which maps a point  576 98 : 8T:UV#W ;<8:=>7?A@B>CEDGF POQ6*R HJILKNM   M (2) Thereby the projection index associates a data point with its nearest neighbour in the . 0 Y to which is parametrized according projection set . 0  ]  X'AQ76 R ;ZR\[P]^1_)a` ! (3) where is the set of allR possible projection indices which are realizations of the deterministic latent variable . For a fixed non-zero kernel bandwidth the parameters of the quantization function may be determined by nonparametric maximum likelihood (ML) estimation, as will be introduced in the next section. For an intuitive motivation of the QDE, one may ask from perspec a data    compression ! for the realization tive whether it is necessary to store all the sample data   of the kernel density estimator or if it is possible to first reduce the data by some suitable quantization method and then construct the estimator from the more parsimonious complexity-reduced data set. Clearly, we would prefer a quantizer which does not decrease the performance of the estimator on unseen and unquantized data. To get an idea of how to select a suitable quantization function let us consider an example from a 1D data space. In one dimension a natural . 0 projection    set   can! be specified by a set of b quantization levels on the real line, i.e. dc  cfe . For a fixed kernel bandwidth, we can now perform maximum likelihood estimation of the level coordinates. In that way we obtain a maximum likelihood estimator of the form  agh5 3j 'mlno?A@BCEDpFZqnj gdrfO e  qsj j  =i  gY c   (4) k c ! counting the number of data points which are with l i quantized to level . In this case, it remains the question how to choose the number of quantization levels. From a different starting point the authors in [3] proposed the same functional form of a nonparametric ML density estimator with respect to Gaussian kernels of equal width cen tered on variable positions. As with the traditional Gaussian KDE (fixed kernel centers on data points), for consistency of the estimator the bandwidth has to be decreased as the sample size increases. In [3] the authors reported that for a fixed non-zero bandwidth, ML estimation of the kernel centers always resulted in a smaller number of actually distinct centers, i.e. several kernels coincided to maximize the   likelihood. Therefore the resulting estimator had the form of (4) where b corresponds to the number of distinct centers with counting the number of kernels coinciding at c . The optimum number of effective i quantization levels for a given bandwidth therefore arises as an automatic byproduct of ML estimation. Finally one has to choose an appropriate kernel width which implicitly determines the complexity of the quantizer. The bandwidth selection problem has been tackled in the domain of kernel density estimation for some time and many approaches have been proposed (see e.g. [5] for an overview), among which the cross-validation methods are most common. In the next section we will adopt the method of likelihood cross-validation to find a practical answer to the bandwidth selection problem. 3 General Learning Scheme By applying the method of sieves as proposed in [3], for a fixed non-zero bandwidth we can   estimate the parameters of the quantization function via maximization of the log-likelihood BZ9a  w.r.t. to . For consistency of the resulting density estimator the band width has to be decreased as the sample size increases, since asymptotically the estimator must converge to a mixture of delta functions centered on the data points. Thus, for decreasing bandwidth, the quantization function of the QDE must converge to the identity function, i.e. the QDE must converge to the kernel density estimator. For a fixed bandwidth, maximization of the likelihood can be achieved by applying the EMalgorithm [2] which provides a convenient optimization scheme, especially for Gaussian kernels. The EM-scheme requires to iterate the following two steps     Yr       ;=q  E-Step: (5)    q        Y    ?A@B5C 0 ?     B   ; r  r M-Step:     (6)    r     s      for a sequence  with suitable initial parameter vector  and sufficient Y r l convergence at   . Thereby i  denotes the posterior probability that ; r  data point has been ?generated? by mixture component k with density . For further insight  one may realize that the M-Step requires to solve a constrained optimization problem by   r  searching for C 0 ?      B   6*R   "! #   R     R     r  Y      R 7?A@B CEDpF  X O 6 R   subject to HJILK M M   r (7) (8) In general this optimization problem can only be solved by iterative techniques. Therefore it may be convenient not to maximize but only to increase the log-likelihood at the M-Step which then corresponds to an application of the generalized EM-algorithm. Without (8) unconstrained maximization according to (7) yields another class of interesting learning schemes which for reasons of space will not be considered in this paper. For Gaussian kernels and an Euclidean metric for the projection, in the limiting case of a vanishing bandwidth, EM-optimization of the QDE parameters corresponds to minimization of the following error or risk  Y      CEDpF  OX6 R J H   ILK M   M Minimization of such error functions corresponds to traditional approaches to unsupervised learning of projection models which can be viewed as a special case of QDE-based learning. 3.1 Bandwidth Selection It is easy to see that the kernel bandwidth cannot be determined by ML-estimation since maximization of the likelihood would drive the bandwidth towards zero. For selection of the kernel bandwidth, we therefore apply the method of likelihood cross-validation (see e.g. [12]), which can be realized by a slight extension of the above EM-scheme. With the  4 leave-one-out QDE      ;Yr  O r (9)      BZ     the idea is to maximize with respect to the kernel bandwidth. For a  Gaussian kernel with bandwidth  an appropriate EM scheme requires the following MStep update rule  Y     r      r  OX r  M   r     M (10)  The posterior probabilities  are easily derived from a leave-one-out version of (5). In an overall optimization scheme one may now alter the estimation of and  or alternatively one may estimate both by likelihood cross-validation. 4 Projection Sets in Multidimensions By the specification of a certain class of quantization functions we can incorporate domain knowledge into the density estimation process, in order to improve generalization. Thereby the idea is to reduce the variance of the density estimator by reducing the variation of the quantized training set. The price is an increase of the bias which requires a careful selection of the set of admissible quantization functions. Then the QDE offers the chance to find a better bias/variance trade-off then with the ?non-quantizing? KDE. We will now show how to utilize existing methods for unsupervised learning within the current density estimation framework. Because many unsupervised methods can be stated in terms of finding optimal projection sets, it is straightforward to apply the corresponding types of quantization functions within the current framework. Thus in the following we shall consider specific parametrizations of the general projection set (3) which correspond to traditional unsupervised learning methods. 4.1 Vector Quantization Vector quantization (VQ) is a standard technique among unsupervised methods and it is easily incorporated into the current density estimation framework by straightforward gen- eralization of the one-dimensional quantizer in section 2 to the multi-dimensional case. Again with a fixed kernel bandwidth ML estimation yields a certain number of b distinct (?effective?) quantization levels, similar to maximum entropy clustering [9, 1]. The projection set of a vector quantizer can be parametrized according to a general basis function representation [7] 4 6  > ;N[     !   $#& with -dimensional vector of basis functions  rL s5 rq for component k . (11) containing discrete delta functions, i.e. The QDE on the basis of a vector quantizer can be expected to generalize well if some cluster structure is present within the data. In multi-dimensional spaces the data are often concentrated in certain regions which allows for a sparse representation by some reference vectors well-positioned in those regions. An alternative approach has been proposed in [14] where the application of the support vector formalism to density estimation results in a sparse representation of the data distribution. 4.2 Principal Component Analysis  projection set yields candidate functions of the form A linear affine parametrization of 4 the 6 R 5 Rs RX[N)a` (12)  with  . The PCA approach reflects our knowledge that in most high-dimensional data spaces, the data are concentrated around some manifold of lower dimensionality. To exploit this structure PCA divides the sample space into two subspaces which are quantized in different ways: within the ?inner? subspace spanned by the directions of the projection manifold we have no quantization at all; within the orthogonal ?outer? subspace the data are quantized to a single level. With a Gaussian kernel with fixed bandwidth  the constrained optimization problem at the M-Step takes a convenient form which facilitates further analysis of the learning algorithm. From (7) and (8) it follows that one has to maximize the following objective function  sN> const. O      r   O   r OL5O   M r M       (13)     matrix has orthogonal where    columns which span the subspace of the projection manifold. From the consideration one finds Q   of the corresponding stationarity conditions  that the sample mean  is an estimator of the shift vector . Maximization of (13) with respect to tr   then requires to maximize the following trace   ! O" O  #a$ S   with symmetric matrices !      r  Thus (14) is maximized   if  values, i.e. with (  (  dimensionality    r  # r   tr &%'    r  r # ;"    r #r     r    (14) (15) % contains all eigenvectors of % , associated with positive eigenwe have the optimal subspace (*) being the eigenvalues of   j  +( ' (  , sal     - ! j (16) which complements a recent result about parametric dimensionality estimation with respect to a  -factor model with isotropic  noise [8]. For the QDE, the two limiting cases   Gaussian of zero and infinite bandwidth,    areO ofparticular   O   interest. With the positive definite sample   covariance matrix one can show DpC  O %   %  DpC    %   (17)  Thus for sufficiently large bandwidth becomes negative definite, which implies a zero   subspace dimensionality estimator  , i.e. all data are quantized to the sample mean. %   For sufficiently small bandwidth becomes positive definite implying  , i.e. no quantization takes place. 4.3 Independent Component Analysis The PCA method provides a rather coarse quantization scheme since it only decides between one-level and no quantization for each subspace dimension. A natural refinement would therefore be to allow for a certain number of effective quantization levels for each component. Such an approach may be viewed as a nonparametric variant of independent component analysis (ICA). The idea is to quantize each coordinate axis separately, which yields a multi-dimensional quantization grid according to   6 R Y    > +         $#&   #s R\[        ! +  (18)  [ )+  [ ) with  , as in (11). Thereby the components of  contain the  l and quantization levels of the -th coordinate axis with direction  . Further, it makes sense to . There are strong similarities with a normalize the direction vectors according to M  M parametric ICA model which has been suggested in [10], where source densities have been mixtures of delta functions and additive noise has been isotropic Gaussian. Other unsupervised learning methods which correspond to different projection sets, like principal curves or multilayer perceptrons (see [7] for an overview) can as well be incorporated into the QDE framework and will be considered elsewhere. 5 Experiments In the following experiments we investigated the PCA based QDE with Gaussian kernel and compared the generalization performance with that of the ?non-quantizing? KDE. All parameters, including the bandwidth of the KDE, were estimated by likelihood crossvalidation. In the first experiment we sampled    h 100 points from a stretched and rotated rectangle. In this case the QDE extracted uniform distribution with support on a a one-dimensional ?unquantized? subspace. Generalization performance was measured by the average log-likelihood on an independent 1000-point test set. With an automatically se lected 1D subspace  (compression ratio ) the PCA-QDE could improve the performance     of the KDE from to . Thus, the PCA-QDE could successfully exploit the elongated structure of the distribution. The estimated density functions are depicted in figure 1,       grid. From the images one can where grey-values are proportional to  on a see, that the QDE better captures the global structure of the distribution while the KDE is more sensitive to local variations in the data. In  a  second experiment we trained PCA-QDEs with -dimensional real-world data ( images) which had been derived from the MNIST database of handwritten digits (http://www.research.att.com/  yann/ocr/mnist/). For each digit class a       -point training set and a  -point test set were used to compare the PCA-QDE with Figure 1: Left: stretched uniform distribution in 2D with white points indicating 100 data points used for estimation; middle: Estimated density using the PCA-QDE; right: kernel density estimate. the KDE, with results shown in table 1. Again the PCA-QDE improved the generalization performance of the KDE although the QDE decided to remove about 40 ?redundant? dimensions per digit class.  Table 1: Results on -dimensional digit data for different digit classes ?0?...?9? (first row); second row: difference between average log-likelihoods of (PCA-)QDE and KDE on test set; third row: optimal subspace dimensionality of QDE Digit:   : : 0 1.87 22 1 0.66 29 2 1.02 26 3 1.38 24 4 1.58 24 5 1.54 25 6 1.44 24 7 0.64 27 8 1.53 21 9 1.33 25 6 Conclusion The QDE offers a nonparametric approach to unsupervised learning of quantization functions which can be viewed as a generalization of the kernel density estimator. While the KDE is directly constructed from the given data set the QDE first creates a quantized representation of the data. Unlike traditional quantization methods which minimize the associated reconstruction error of the data points, the QDE adjusts the quantizer to optimize an estimate of the data density. This feature allows for a convenient model selection procedure, since the complexity of the quantizer can be controlled by the kernel bandwidth, which in turn can be selected in a data-driven way. For a practical realization we outlined EM-schemes for parameter estimation and bandwidth selection. As an illustration, we discussed examples with different projection sets which correspond to VQ, PCA and ICA methods. We presented experiments which demonstrate that the bias imposed by the quantization can lead to an improved generalization as compared to the ?non-quantizing? KDE. This suggests that QDEs offer a promising approach to unsupervised learning that allows to control bias without the usually rather strong distributional assumptions of the Bayesian approach. Acknowledgement This work was funded by the Deutsche Forschungsgemeinschaft within the project SFB 360. References [1] J. M. Buhmann and N. Tishby. Empirical risk approximation: A statistical learning theory of data clustering. In C. M. Bishop, editor, Neural Networks and Machine Learning, pages 57?68. Springer, Berlin Heidelberg New York, 1998. [2] A. P. Dempster, N. M. Laird, and D. B. Rubin. Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society Series B, 39:1?38, 1977. [3] Stuart Geman and Chii-Ruey Hwang. Nonparametric maximum likelihood estimation by the method of sieves. The Annals of Statistics, 10(2):401?414, 1982. [4] T. Hastie and W. Stuetzle. Principal curves. Journal of the American Statistical Association, 84:502?516, 1989. [5] M. C. Jones, J. S. Marron, and S. J. Sheather. A brief survey of bandwidth selection for density estimation. Journal of the American Statistical Association, 91(433):401? 407, 1996. [6] B. K?egl, A. Krzyzak, T. Linder, and K. Zeger. Learning and design of principal curves. IEEE Transaction on Pattern Analysis and Machine Intelligence, 22(3):281? 297, 2000. [7] Peter Meinicke. Unsupervised Learning in a Generalized Regression Framework. PhD thesis, Universitaet Bielefeld, 2000. http://archiv.ub.unibielefeld.de/disshabi/2000/0033/. [8] Peter Meinicke and Helge Ritter. Resolution-based complexity control for Gaussian mixture models. Neural Computation, 13(2):453?475, 2001. [9] K. Rose, E. Gurewitz, and G. C. Fox. Statistical mechanics and phase transitions in clustering. Physical Review Letters, 65(8):945?948, 1990. [10] Sam Roweis and Zoubin Ghahramani. A unifying review of linear Gaussian models. Neural Computation, 11(2):305?345, 1999. [11] D. W. Scott. Multivariate Density Estimation. Wiley, 1992. [12] B. W. Silverman. Density Estimation for Statistics and Data Analysis. Chapman and Hall, London and New York, 1986. [13] Alex J. Smola, Robert C. Williamson, Sebastian Mika, and Bernhard Sch?olkopf. Regularized principal manifolds. In Proc. 4th European Conference on Computational Learning Theory, volume 1572, pages 214?229. Springer-Verlag, 1999. [14] Vladimir N. Vapnik and Sayan Mukherjee. Support vector method for multivariate density estimation. In S. A. Solla, T. K. Leen, and K.-R. M?uller, editors, Advances in Neural Information Processing Systems, volume 12, pages 659?665. The MIT Press, 2000.
2007 |@word middle:1 version:1 compression:2 meinicke:3 grey:1 covariance:1 thereby:4 tr:2 reduction:1 initial:1 contains:1 att:1 series:1 existing:1 current:3 com:1 must:3 realize:1 additive:1 zeger:1 mstep:1 remove:1 update:1 implying:1 generative:2 selected:1 yr:2 intelligence:1 mln:1 isotropic:2 parametrization:1 vanishing:1 quantizer:9 quantized:9 provides:2 coarse:1 constructed:1 ica:3 expected:1 mechanic:1 multi:3 unquantized:2 decreasing:1 automatically:1 becomes:2 project:1 underlying:3 deutsche:1 kind:1 interpreted:1 finding:2 tackle:1 control:2 appear:1 positive:3 local:1 mika:1 suggests:1 decided:1 practical:3 definite:3 silverman:1 digit:6 stuetzle:1 procedure:1 empirical:1 significantly:1 projection:19 convenient:4 suggest:1 zoubin:1 get:1 cannot:2 unlabeled:1 selection:11 risk:2 applying:2 optimize:1 www:1 map:1 imposed:1 center:4 elongated:1 deterministic:1 straightforward:2 starting:1 survey:1 resolution:1 estimator:22 insight:1 rule:1 adjusts:1 spanned:1 searching:1 coordinate:3 variation:2 limiting:2 techfak:2 construction:2 target:1 annals:1 hjilk:1 associate:1 areo:1 mukherjee:1 distributional:1 database:1 geman:1 solved:1 capture:1 region:2 solla:1 decrease:1 trade:1 rose:1 dempster:1 complexity:6 trained:1 creates:1 basis:4 easily:2 distinct:3 effective:3 london:1 neuroinformatics:2 solve:1 reconstruct:2 statistic:2 unseen:1 laird:1 sequence:1 eigenvalue:1 quantizing:6 reconstruction:2 realization:3 gen:1 parametrizations:1 achieve:1 roweis:1 intuitive:1 normalize:1 olkopf:1 crossvalidation:1 convergence:1 cluster:1 optimum:1 pmeinick:1 leave:2 rotated:1 measured:1 nearest:1 strong:3 implies:1 direction:3 centered:1 human:1 generalization:9 extension:1 around:1 considered:2 sufficiently:2 hall:1 cen:1 adopt:1 lected:1 estimation:26 proc:1 sensitive:1 successfully:1 reflects:1 minimization:2 uller:1 mit:1 clearly:1 gaussian:11 always:2 aim:1 rather:3 derived:3 likelihood:16 contrast:1 sense:1 germany:2 overall:1 among:2 constrained:2 special:1 equal:1 construct:2 chapman:1 stuart:1 jones:1 unsupervised:18 alter:1 modern:1 neighbour:1 resulted:1 phase:1 multidimensions:1 stationarity:1 interest:1 mixture:4 byproduct:1 necessary:1 orthogonal:2 fox:1 incomplete:1 euclidean:1 divide:1 eralization:1 formalism:1 column:1 maximization:5 subset:1 uniform:2 tishby:1 optimally:2 reported:1 answer:1 marron:1 synthetic:1 density:34 ritter:2 off:1 again:2 reflect:1 thesis:1 containing:1 choose:2 american:2 de:3 gy:1 try:1 minimize:1 variance:2 maximized:1 correspond:4 yield:4 generalize:1 chii:1 bayesian:2 handwritten:1 rx:1 drive:1 explain:1 sebastian:1 associated:2 tered:1 sampled:1 popular:1 ask:1 knowledge:3 dimensionality:5 positioned:1 actually:1 specify:1 coinciding:1 improved:2 leen:1 ox:1 smola:1 hwang:1 building:1 contain:1 sieve:2 symmetric:1 white:1 width:3 generalized:2 demonstrate:1 gh:1 image:2 consideration:2 common:3 functional:2 rl:1 overview:3 physical:1 volume:2 discussed:1 association:2 slight:1 stretched:2 automatic:1 unconstrained:1 outlined:1 grid:2 consistency:2 had:2 funded:1 specification:1 similarity:1 surface:1 posterior:2 multivariate:2 recent:1 driven:3 store:1 certain:5 verlag:1 converge:3 maximize:5 redundant:1 cross:4 offer:3 controlled:1 variant:1 basic:1 regression:1 multilayer:1 essentially:1 metric:1 bz:2 kernel:27 sof:1 achieved:1 separately:1 addressed:1 decreased:2 source:1 sch:1 unlike:1 subject:1 facilitates:1 sheather:1 counting:2 forschungsgemeinschaft:1 easy:1 concerned:1 iterate:1 hastie:1 bandwidth:26 reduce:2 idea:4 inner:1 shift:1 whether:1 pca:14 sfb:1 krzyzak:1 peter:3 york:2 se:1 eigenvectors:1 nonparametric:9 band:1 concentrated:2 reduced:2 http:2 delta:3 estimated:3 per:1 discrete:1 shall:1 group:2 redundancy:1 utilize:2 rectangle:1 asymptotically:1 realworld:1 letter:1 place:1 bielefeld:7 almost:1 yann:1 parsimonious:1 prefer:1 tackled:1 ilk:1 alex:1 archiv:1 span:1 ruey:1 according:5 smaller:1 em:7 sam:1 helge:3 vq:3 remains:1 turn:1 mechanism:1 apply:2 ocr:1 appropriate:2 alternative:3 denotes:2 clustering:3 unifying:1 const:1 exploit:3 ghahramani:1 especially:1 society:1 objective:3 question:1 realized:1 parametric:2 traditional:5 subspace:10 berlin:1 parametrized:4 outer:1 manifold:4 reason:1 index:3 illustration:1 ratio:1 vladimir:1 unfortunately:1 robert:1 kde:12 trace:1 stated:1 negative:1 design:1 perform:1 finite:1 incorporated:2 dc:1 overlooked:1 inferred:1 introduced:1 tive:1 namely:1 complement:1 specified:1 suggested:1 usually:5 pattern:1 scott:1 including:1 royal:1 suitable:5 natural:2 regularized:1 buhmann:1 zr:1 dpc:2 scheme:9 improve:3 brief:1 axis:2 gurewitz:1 review:2 acknowledgement:1 interesting:1 proportional:1 validation:4 affine:1 sufficient:1 rubin:1 editor:2 row:3 elsewhere:1 cfe:1 bias:4 allow:1 sparse:2 curve:4 dimension:3 world:1 transition:1 author:2 made:1 refinement:1 transaction:1 uni:2 implicitly:1 bernhard:1 ml:6 global:1 decides:1 universitaet:1 alternatively:1 latent:1 iterative:1 table:2 promising:1 heidelberg:1 quantize:1 williamson:1 investigated:1 european:1 domain:3 motivation:1 noise:2 referred:1 wiley:1 position:1 candidate:1 third:1 coincided:1 admissible:1 sayan:1 specific:4 bishop:1 quantization:33 mnist:2 vapnik:1 phd:1 egl:1 entropy:1 depicted:1 springer:2 corresponds:4 determines:1 chance:1 extracted:1 viewed:4 identity:1 careful:1 towards:1 price:1 determined:2 infinite:1 reducing:1 principal:8 perceptrons:1 indicating:1 select:2 linder:1 support:3 latter:1 arises:1 ub:1 incorporate:1
1,105
2,008
Model Based Population Tracking and Automatic Detection of Distribution Changes Igor V. Cadez ? Dept. of Information and Computer Science, University of California, Irvine, CA 92612 icadez@ics.uci.edu P. S. Bradley digiMine, Inc. 10500 NE 8th Street, Bellevue, WA 98004-4332 paulb@digimine.com Abstract Probabilistic mixture models are used for a broad range of data analysis tasks such as clustering, classification, predictive modeling, etc. Due to their inherent probabilistic nature, mixture models can easily be combined with other probabilistic or non-probabilistic techniques thus forming more complex data analysis systems. In the case of online data (where there is a stream of data available) models can be constantly updated to reflect the most current distribution of the incoming data. However, in many business applications the models themselves represent a parsimonious summary of the data and therefore it is not desirable to change models frequently, much less with every new data point. In such a framework it becomes crucial to track the applicability of the mixture model and detect the point in time when the model fails to adequately represent the data. In this paper we formulate the problem of change detection and propose a principled solution. Empirical results over both synthetic and real-life data sets are presented. 1 Introduction and Notation Consider a data set D = {x1 , x2 , . . . , xn } consisting of n independent, identically distributed (iid) data points. In context of this paper the data points could be vectors, sequences, etc. Further, consider a probabilistic mixture model that maps each data set to a real number, the probability of observing the data set: P (D|?) = n Y i=1 P (xi |?) = n X K Y ?k P (xi |?k ), (1) i=1 k=1 where the model is parameterized by ? = {?1 , . . . , ?K , ?1 , . . . , ?K }. Each P (.|?k ) represents a mixture component, while ?i represents mixture weights. It is often more convenient ? Work was done while author was at digiMine, Inc., Bellevue, WA. to operate with the log of the probability and define the log-likelihood function as: l(?|D) = log P (D|?) = n X log P (xi |?) = i=1 n X LogPi i=1 which is additive over data points rather than multiplicative. The LogPi terms we introduce in the notation represent each data point?s contribution to the overall log-likelihood and therefore describe how well a data point fits under the model. For example, Figure 3 shows a distribution of LogP scores using a mixture of conditionally independent (CI) models. Maximizing probability1 of the data with respect to the parameters ? can be accomplished by the Expectation-Maximization (EM) algorithm [6] in linear time in both data complexity (e.g., number of dimensions) and data set size (e.g., number of data points). Although EM guarantees only local optimality, it is a preferred method for finding good solutions in linear time. We consider an arbitrary but fixed parametric form of the model, therefore we sometimes refer to a specific set of parameters ? as the model. Note that since the logarithm is a monotonic function, the optimal set of parameters is the same whether we use likelihood or log-likelihood. Consider an online data source where there are data sets Dt available at certain time intervals t (not necessarily equal time periods or number of data points). For example, there could be a data set generated on a daily basis, or it could represent a constant stream of data from a monitoring device. In addition, we assume that we have an initial model ?0 that was built (optimized, fitted) on some in-sample data D0 = {D1 , D2 , . . . , Dt0 }. We would like to be able to detect a change in the underlying distribution of data points within data sets Dt that would be sufficient to require building of a new model ?1 . The criterion for building a new model is loosely defined as ?the model does not adequately fit the data anymore?. 2 Model Based Population Similarity In this section we formulate the problem of model-based population similarity and tracking. In case of mixture models we start with the following observations: ? The mixture model defines the probability density function (PDF) that is used to score each data point (LogP scores), leading to the score for the overall population (log-likelihood or sum of LogP scores). ? The optimal mixture model puts more PDF mass over dense regions in the data space. Different components allow the mixture model to distribute its PDF over disconnected dense regions in the data space. More PDF mass in a portion of the data space implies higher LogP scores for the data points lying in that region of the space. ? If model is to generalize well (e.g., there is no significant overfitting) it cannot put significant PDF mass over regions of data space that are populated by data points solely due to the details of a specific data sample used to build the model. ? Dense regions in the data space discovered by a non-overfitting model are the intrinsic property of the true data-generating distribution even if the functional form of the model is not well matched with the true data generating distribution. In the latter case, the model might not be able to discover all dense regions or might not model the correct shape of the regions, but the regions that are discovered (if any) are intrinsic to the data. 1 This approach is called maximum-likelihood estimation. If we included parameter priors we could equally well apply results in this paper to the maximum a posteriori estimation. ? If there is confidence that the model is not overfitting and that it generalizes well (e.g., cross-validation was used to determine the optimal number of mixture components), the new data from the same distribution as the in-sample data should be dense in the same regions that are predicted by the model. Given these observations, we seek to define a measure of data-distribution similarity based on how well the dense regions of the data space are preserved when new data is introduced. In model based clustering, dense regions are equivalent to higher LogP scores, hence we cast the problem of determining data distribution similarity into one of determining LogP distribution similarity (relative to the model). For example, Figure 3 (left) shows a histogram of one such distribution. It is important to note several properties of Figure 3: 1) there are several distinct peaks from which distribution tails off toward smaller LogP values, therefore simple summary scores fail to efficiently summarize the LogP distribution. For example, log-likelihood is proportional to the mean of LogP distribution in Figure 3, and the mean is not a very useful statistic when describing such a multimodal distribution (also confirmed experimentally); 2) the histogram itself is not a truly non-parametric representation of the underlying distribution, given that the results are dependent on bin width. In passing we also note that the shape of the histogram in Figure 3 is a consequence of the CI model we use: different peaks come from different discrete attributes, while the tails come from continuous Gaussians. It is a simple exercise to show that LogP scores for a 1-dimensional data set generated by a single Gaussian have an exponential distribution with a sharp cutoff on the right and tail toward the left. To define the similarity of the data distributions based on LogP scores in a purely nonparametric way we have at our disposal the powerful formalism of Kolmogorov-Smirnov (KS) statistics [7]. KS statistics make use of empirical cumulative distribution functions (CDF) to estimate distance between two empirical 1-dimensional distributions, in our case distributions of LogP scores. In principle, we could compare the LogP distribution of the new data set Dt to that of the training set D0 and obtain the probability that the two came from the same distribution. In practice, however, this approach is not feasible since we do not assume that the estimated model and the true data generating process share the same functional form (see Section 3). Consequently, we need to consider the specific KS score in relation to the natural variability of the true data generating distribution. In the situation with streaming data, the model is estimated over the in-sample data D0 . Then the individual in-sample data sets D1 , D2 , . . . , Dt0 are used to estimate the natural variability of the KS statistics. This variability needs to be quantified due to the fact that the model may not truly match the data distribution. When the natural variance of the KS statistics over the in-sample data has been determined, the LogP scores for a new dataset Dt , t > t0 are computed. Using principled heuristics, one can then determine whether or not the LogP signature for Dt is significantly different than the LogP signatures for the in-sample data. To clarify various steps, we provide an algorithmic description of the change detection process. Algorithm 1 (Quantifying Natural Variance of KS Statistics): Given an ?in-sample? dataset D0 = {D1 , D2 , . . . , Dt0 }, proceed as follows: 1. Estimate the parameters ?0 of the mixture model P (D|?) over D0 (see equation (1)). 2. Compute ni X (2) LogP (Di ) = log P (x?i |?0 ), x?i ? Di , ni = |Di |, i = 1, . . . , t0 . ?i=1 3. For 1 ? i, j ? t0 , compute LKS (i, j) = log [PKS (Di , Dj )]. See [7] for details on PKS computation. 4. For 1 ? i ? t0 , compute the KS measure MKS (i) as Pt 0 j=1 LKS (i, j) . MKS (i) = t0 (3) 5. Compute ?M = M ean[MKS (i)] and ?M = ST D[MKS (i)] to quantify the natural variability of MKS over the ?in-sample? data. Algorithm 2 (Evaluating New Data): Given a new dataset Dt , t > t0 , ?M and ?M proceed as follows: 1. 2. 3. 4. Compute LogP (Dt ) as in (2). For 1 ? i ? t0 , compute LKS (i, t). Compute MKS (t) as in (3). Apply decision criteria using MKS (t), ?M , ?M to determine whether or not ?0 is a good fit for the new data. For example, if |MKS (t) ? ?M | > 3, ?M then ?0 is not a good fit any more. (4) Note, however, that the 3-? interval be interpreted as a confidence interval only in the limit when number of data sets goes to infinity. In applications presented in this paper we certainly do not have that condition satisfied and we consider this approach as an ?educated heuristic? (gaining firm statistical grounds in the limit). 2.1 Space and Time Complexity of the Methodology The proposed methodology was motivated by a business application with large data sets, hence it must have time complexity that is close to linear in order to scale well. In order to assess the time complexity, we use the following notation: nt = |Dt | is the number of data points in the data set Dt ; t0 is the index ofP the last in-sample data set, but is also the t0 number of in-sample data sets; n0 = |D0 | = t=1 nt is the total number of in-sample data points (in all the in-sample data sets); n = n0 /t0 is the average number of data points in the in-sample data sets. For simplicity of argument, we assume that all the data sets are approximately of the same size, that is nt ? n. The analysis presented here does not take into account the time and space complexity needed to estimated the parameters ? of the mixture model (1). In the first phase of the methodology, we must score each of the in-sample data points under the model (to obtain the LogP distributions) which has time complexity of O(n0 ). Calculation of KS statistics for two data sets is done in one pass over the LogP distributions, but it requires that the LogP scores be sorted, hence it has time complexity of 2n + 2O(n log n) = O(n log n). Since we must calculate all the pairwise KS measures, this step has time complexity of t0 (t0 ? 1)/2 O(n log n) = O(t20 n log n). In-sample mean and variance of the KS measure are obtained in time which is linear in t0 hence the asymptotic time complexity does not change. In order to evaluate out-of-sample data sets we must keep LogP distributions for each of the in-sample data sets as well as several scalars (e.g., mean and variance of the in-sample KS measure) which requires O(n0 ) memory. To score an out-of-sample data set Dt , t > t0 , we must first obtain the LogP distribution of Dt which has time complexity of O(n) and then calculate the KS measure relative to each of the in-sample data sets which has time complexity O(n log n) per in-sample data set, or t0 O(n log n) = O(t0 n log n) for the full in-sample period. The LogP distribution for Dt can be discarded once the pairwise KS measures are obtained. 3000 3000 2500 2500 2000 2000 Count 3500 Count 3500 1500 1500 1000 1000 500 500 0 ?5.5 ?5 ?4.5 ?4 ?3.5 ?3 0 ?2.5 ?5.5 ?5 ?4.5 LogP ?4 ?3.5 ?3 ?2.5 ?4 ?3.5 ?3 ?2.5 LogP 3000 3000 2500 2500 2000 2000 Count 3500 Count 3500 1500 1500 1000 1000 500 500 0 ?5.5 ?5 ?4.5 ?4 ?3.5 ?3 LogP ?2.5 0 ?5.5 ?5 ?4.5 LogP Figure 1: Histograms of LogP scores for two data sets generated from the first model (top row) and two data sets generated from the second model (bottom row). Each data set contains 50,000 data points. All histograms are obtained from the model fitted on the in-sample period. Overall, the proposed methodology requires O(n0 ) memory, O(t20 n log n) time for preprocessing and O(t0 n log n) time for out-of-sample evaluation. Further, since t0 is typically a small constant (e.g., t0 = 7 or t0 = 30), the out-of-sample evaluation practically has time complexity of O(n log n). 3 Experimental Setup Experiments presented consist of two parts: experiments on synthetic data and experiments on the aggregations over real web-log data. 3.1 Experiments on Synthetic Data Synthetic data is a valuable tool when determining both applicability and limitations of the proposed approach. Synthetic data was generated by sampling from a a two component CI model (the true model is not used in evaluations). The data consist of a two-state discrete dimension and a continuous dimension. First 100 data sets where generated by sampling from a mixture model with parameters: [?1 , ?2 ] = [0.6, 0.4] as weights, ?1 = [0.8, 0.2] and ?2 = [0.4, 0.6] as discrete state probabilities, [?1 , ?12 ] = [10, 5] and [?2 , ?22 ] = [0, 7] as mean and variance (Gaussian) for the continuous variable. Then the discrete dimension probability of the second cluster was changed from ?2 = [0.4, 0.6] to ?0 2 = [0.5, 0.5] keeping the remaining parameters fixed and an additional 100 data sets were generated by sampling from this altered model. This is a fairly small change in the distribution and the underlying LogP scores appear to be very similar as can be seen in Figure 1. The figure shows LogP distributions for the first two data sets generated from the first model (top row) and the first two data sets generated from the second model (bottom row). Plots within each 0 0 ?1 ?5 <log(KS probability)> <log(KS probability)> ?2 ?3 ?4 ?10 ?5 (b) (a) ?6 0 20 40 60 80 100 Data set Dt 120 140 160 180 ?15 200 0 0 20 40 60 80 100 Data set Dt 120 140 160 180 200 40 60 80 100 Data set Dt 120 140 160 180 200 0 ?5 ?2 ?10 ?15 <log(KS probability)> <log(KS probability)> ?4 ?6 ?8 ?20 ?25 ?30 ?35 ?10 ?40 ?12 ?45 (c) ?14 0 20 40 60 80 100 Data set Dt 120 140 160 180 200 ?50 (d) 0 20 Figure 2: Average log(KS probability) over the in-sample period for four experiments on synthetic data, varying the number of data points per data set: a) 1,000; b) 5,000; c) 10,000; d) 50,000. The dotted vertical line separates in-sample and out-of-sample periods. Note that y-axes have different scales in order to show full variability of the data. row should be more similar than plots from different rows, but this is difficult to discern by visual inspection. Algorithms 1 and 2 were evaluated by using the first 10 data sets to estimate a two component model. Then pairwise KS measures were calculated between all possible data set pairs relative to the estimated model. Figure 2 shows average KS measures over in-sample data sets (first 10) for four experiments varying the number of data points in each experiment. Note that the vertical axes are different in each of the plots to better show the range of values. As the number of data points in the data set increases, the change that occurs at t = 101 becomes more apparent. At 50,000 data points (bottom right plot of Figure 2) the change in the distribution becomes easily detectable. Since this number of data points is typically considered to be small compared to the number of data points in our real life applications we expect to be able to detect such slight distribution changes. 3.2 Experiments on Real Life Data Figure 3 shows a distribution for a typical day from a content web-site. There are almost 50,000 data points in the data set with over 100 dimensions each. The LogP score distribution is similar to that of synthetic data in Figure 1 which is a consequence of the CI model used. Note, however, that in this data set the true generating distribution is not known and is unlikely to be purely a CI model. Therefore, the average log KS measure over insample data has much lower values (see Figure 3 right, and plots in Figure 2). Another way to phrase this observation is to note that since the true generating data distribution is most likely not CI, the observed similarity of LogP distributions (the KS measure) is much lower since there are two factors of dissimilarity: 1) different data sets; 2) inability of the CI model to capture all the aspects of the true data distribution. Nonetheless, the first 31 ?100 5000 ?200 4500 4000 ?300 <log(KS probability)> 3500 Count 3000 2500 2000 ?400 ?500 ?600 1500 1000 ?700 500 0 ?100 ?800 ?80 ?60 ?40 ?20 LogP 0 20 40 60 0 10 20 30 40 50 Data set D 60 70 80 90 100 t Figure 3: Left: distribution of 42655 LogP scores from mixture of conditional independence models. The data is a single-day of click-stream data from a commercial web site. Right: Average log(KS probability) over the 31 day in-sample period for a content website showing a glitch on day 27 and a permanent change on day 43, both detected by the proposed methodology. data sets (one month of data) that were used to build the initial model ?0 can be used to define the natural variability of the KS measures against which additional data sets can be compared. The result is that in Figure 3 we clearly see a problem with the distribution on day 27 (a glitch in the data) and a permanent change in the distribution on day 43. Both of the detected changes correspond to real changes in the data, as verified by the commercial website operators. Automatic description of changes in the distribution and criteria for automatic rebuilding of the model are beyond scope of this paper. 4 Related Work Automatic detection of various types of data changes appear in the literature in several different flavors. For example, novelty detection ([4], [8]) is the task of determining unusual or novel data points relative to some model. This is closely related to the outlier detection problem ([1], [5]) where the goal is not only to find unusual data points, but the ones that appear not to have been generated by the data generating distribution. A related problem has been addressed by [2] in the context of time series modeling where outliers and trends can contaminate the model estimation. More recently mixture models have been applied more directly to outlier detection [3]. The method proposed in this paper addesses a different problem. We are not interested in new and unusual data points; on the contrary, the method is quite robust with respect to outliers. An outlier or two do not necessarily mean that the underlying data distribution has changed. Also, some of the distribution changes we are interested in detecting might be considered uninteresting and/or not-novel; for example, a slight shift of the population as a whole is something that we certainly detect as a change but it is rarely considered novel unless the shift is drastic. There is also a set of online learning algorithms that update model parameters as the new data becomes available (for variants and additional references, e.g. [6]). In that framework there is no such concept as a data distribution change since the models are constantly updated to reflect the most current distribution. For example, instead of detecting a slight shift of the population as a whole, online learning algorithms update the model to reflect the shift. 5 Conclusions In this paper we introduced a model-based method for automatic distribution change detection in an online data environment. Given the LogP distribution data signature we further showed how to compare different data sets relative to the model using KS statistics and how to obtain a single measure of similarity between the new data and the model. Finally, we discussed heuristics for change detection that become principled in the limit as the number of possible data sets increases. Experimental results over synthetic and real online data indicate that the proposed methodology is able to alert the analyst to slight distributional changes. This methodology may be used as the basis of a system to automatically re-estimate parameters of a mixture model on an ? as-needed? basis ? when the model fails to adequately represent the data after a certain point in time. References [1] V. Barnett and T. Lewis. Outliers in statistical data. Wiley, 1984. [2] A. G. Bruce, J. T. Conor, and R. D. Martin. Prediction with robustness towards outliers, trends, and level shifts. In Proceedings of the Third International Conference on Neural Networks in Financial Engineering, pages 564?577, 1996. [3] I. V. Cadez, P. Smyth, and H. Mannila. Probabilistic modeling of transaction data with applications to profiling, visualization, and prediction. In F. Provost and R. Srikant, editors, Proceedings of the Seventh ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 37?46. ACM, 2001. [4] C. Campbell and K. P. Bennett. A linear programming approach to novelty detection. In T. K. Leen, T. G. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing Systems 13, pages 395?401. MIT Press, 2001. [5] T. Fawcett and F. J. Provost. Activity monitoring: Noticing interesting changes in behavior. In Proceedings of the Fifth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 53?62, 1999. [6] R. Neal and G. Hinton. A view of the em algorithm that justifies incremental, sparse and other variants. In M. I. Jordan, editor, Learning in Graphical Models, pages 355?368. Kluwer Academic Publishers, 1998. [7] W. H. Press, S. A. Teukolsky, W. T. Vetterling, and B. P. Flannery. Numerical Recipes in C: The Art of Scientific Computing, Second Edition. Cambridge University Press, Cambridge, UK, 1992. [8] B. Sch?olkopf, R. C. Williamson, A. J. Smola, J. Shawe-Taylor, and J. C. Platt. Support vector method for novelty detection. In S. A. Solla, T. K. Leen, and K.-R. Mller, editors, Advances in Neural Information Processing Systems 12, pages 582?588. MIT Press, 2000.
2008 |@word smirnov:1 d2:3 seek:1 bellevue:2 initial:2 contains:1 score:20 series:1 cadez:2 bradley:1 current:2 com:1 nt:3 must:5 numerical:1 additive:1 shape:2 plot:5 update:2 n0:5 device:1 website:2 inspection:1 detecting:2 insample:1 alert:1 become:1 introduce:1 pairwise:3 behavior:1 themselves:1 frequently:1 automatically:1 becomes:4 discover:1 notation:3 underlying:4 matched:1 mass:3 interpreted:1 finding:1 guarantee:1 contaminate:1 every:1 uk:1 platt:1 appear:3 educated:1 engineering:1 local:1 limit:3 consequence:2 solely:1 approximately:1 might:3 k:26 quantified:1 range:2 practice:1 mannila:1 empirical:3 significantly:1 convenient:1 confidence:2 cannot:1 close:1 operator:1 put:2 context:2 equivalent:1 map:1 logpi:2 maximizing:1 go:1 formulate:2 simplicity:1 financial:1 population:6 updated:2 pt:1 commercial:2 smyth:1 programming:1 probability1:1 trend:2 distributional:1 bottom:3 observed:1 capture:1 calculate:2 region:11 solla:1 valuable:1 principled:3 environment:1 complexity:12 signature:3 predictive:1 purely:2 basis:3 easily:2 multimodal:1 various:2 kolmogorov:1 distinct:1 describe:1 detected:2 rebuilding:1 firm:1 dt0:3 heuristic:3 apparent:1 quite:1 statistic:8 itself:1 online:6 sequence:1 propose:1 mks:8 uci:1 description:2 olkopf:1 recipe:1 cluster:1 generating:7 incremental:1 predicted:1 implies:1 come:2 quantify:1 indicate:1 closely:1 correct:1 attribute:1 bin:1 require:1 clarify:1 lying:1 practically:1 considered:3 ic:1 ground:1 algorithmic:1 scope:1 mller:1 estimation:3 tool:1 mit:2 clearly:1 gaussian:2 rather:1 varying:2 ax:2 likelihood:7 sigkdd:2 detect:4 posteriori:1 dependent:1 streaming:1 vetterling:1 typically:2 unlikely:1 relation:1 interested:2 overall:3 classification:1 art:1 fairly:1 equal:1 once:1 sampling:3 barnett:1 represents:2 broad:1 igor:1 inherent:1 individual:1 phase:1 consisting:1 detection:11 mining:2 evaluation:3 certainly:2 mixture:18 truly:2 daily:1 unless:1 loosely:1 logarithm:1 taylor:1 re:1 fitted:2 formalism:1 modeling:3 conor:1 logp:36 maximization:1 phrase:1 applicability:2 uninteresting:1 seventh:1 synthetic:8 combined:1 st:1 density:1 peak:2 international:3 probabilistic:6 off:1 reflect:3 satisfied:1 leading:1 account:1 distribute:1 inc:2 permanent:2 stream:3 multiplicative:1 view:1 observing:1 portion:1 start:1 aggregation:1 bruce:1 contribution:1 ass:1 ni:2 variance:5 efficiently:1 correspond:1 generalize:1 iid:1 monitoring:2 confirmed:1 against:1 nonetheless:1 di:4 irvine:1 dataset:3 knowledge:2 campbell:1 disposal:1 higher:2 dt:16 day:7 methodology:7 leen:2 done:2 evaluated:1 smola:1 web:3 defines:1 scientific:1 building:2 dietterich:1 concept:1 true:8 adequately:3 hence:4 neal:1 conditionally:1 width:1 criterion:3 pdf:5 novel:3 recently:1 functional:2 tail:3 slight:4 discussed:1 kluwer:1 refer:1 significant:2 cambridge:2 automatic:5 populated:1 shawe:1 dj:1 similarity:8 etc:2 something:1 showed:1 certain:2 came:1 life:3 accomplished:1 seen:1 additional:3 determine:3 novelty:3 period:6 full:2 desirable:1 d0:6 match:1 academic:1 calculation:1 cross:1 profiling:1 dept:1 ofp:1 equally:1 prediction:2 variant:2 expectation:1 histogram:5 represent:5 sometimes:1 fawcett:1 preserved:1 addition:1 interval:3 addressed:1 source:1 crucial:1 publisher:1 sch:1 operate:1 contrary:1 jordan:1 identically:1 independence:1 fit:4 click:1 shift:5 t0:20 whether:3 motivated:1 passing:1 proceed:2 useful:1 nonparametric:1 pks:2 srikant:1 dotted:1 estimated:4 track:1 per:2 discrete:4 four:2 cutoff:1 verified:1 sum:1 parameterized:1 powerful:1 noticing:1 discern:1 almost:1 parsimonious:1 decision:1 activity:1 infinity:1 x2:1 aspect:1 argument:1 optimality:1 martin:1 disconnected:1 smaller:1 em:3 outlier:7 equation:1 visualization:1 describing:1 count:5 fail:1 detectable:1 needed:2 drastic:1 unusual:3 generalizes:1 available:3 gaussians:1 apply:2 anymore:1 robustness:1 top:2 clustering:2 remaining:1 graphical:1 build:2 occurs:1 parametric:2 distance:1 separate:1 street:1 toward:2 analyst:1 index:1 glitch:2 setup:1 difficult:1 vertical:2 observation:3 discarded:1 situation:1 hinton:1 variability:6 discovered:2 provost:2 arbitrary:1 sharp:1 introduced:2 cast:1 pair:1 optimized:1 california:1 able:4 beyond:1 summarize:1 built:1 gaining:1 t20:2 memory:2 business:2 natural:6 altered:1 ne:1 tresp:1 prior:1 literature:1 discovery:2 determining:4 relative:5 asymptotic:1 expect:1 interesting:1 limitation:1 proportional:1 validation:1 sufficient:1 principle:1 editor:4 share:1 row:6 summary:2 changed:2 last:1 keeping:1 allow:1 fifth:1 sparse:1 distributed:1 dimension:5 xn:1 evaluating:1 cumulative:1 calculated:1 author:1 preprocessing:1 transaction:1 preferred:1 keep:1 overfitting:3 incoming:1 xi:3 continuous:3 nature:1 lks:3 robust:1 ca:1 ean:1 williamson:1 complex:1 necessarily:2 dense:7 whole:2 edition:1 x1:1 site:2 wiley:1 fails:2 exponential:1 exercise:1 third:1 specific:3 showing:1 intrinsic:2 consist:2 ci:7 dissimilarity:1 justifies:1 flavor:1 flannery:1 likely:1 forming:1 visual:1 tracking:2 scalar:1 monotonic:1 constantly:2 lewis:1 cdf:1 acm:3 teukolsky:1 conditional:1 sorted:1 month:1 goal:1 consequently:1 quantifying:1 towards:1 bennett:1 feasible:1 change:23 experimentally:1 included:1 determined:1 typical:1 content:2 called:1 total:1 pas:1 experimental:2 rarely:1 support:1 latter:1 inability:1 evaluate:1 d1:3
1,106
2,009
MIME: Mutual Information Minimization and Entropy Maximization for Bayesian Belief Propagation Anand Rangarajan Dept. of Computer and Information Science and Engineering University of Florida Gainesville, FL 32611-6120, US anand@cise.ufl.edu Alan L. Yuille Smith-Kettlewell Eye Research Institute 2318 Fillmore St. San Francisco, CA 94115, US yuille@ski.org Abstract Bayesian belief propagation in graphical models has been recently shown to have very close ties to inference methods based in statistical physics. After Yedidia et al. demonstrated that belief propagation fixed points correspond to extrema of the so-called Bethe free energy, Yuille derived a double loop algorithm that is guaranteed to converge to a local minimum of the Bethe free energy. Yuille?s algorithm is based on a certain decomposition of the Bethe free energy and he mentions that other decompositions are possible and may even be fruitful. In the present work, we begin with the Bethe free energy and show that it has a principled interpretation as pairwise mutual information minimization and marginal entropy maximization (MIME). Next, we construct a family of free energy functions from a spectrum of decompositions of the original Bethe free energy. For each free energy in this family, we develop a new algorithm that is guaranteed to converge to a local minimum. Preliminary computer simulations are in agreement with this theoretical development. 1 Introduction In graphical models, Bayesian belief propagation (BBP) algorithms often (but not always) yield reasonable estimates of the marginal probabilities at each node [6]. Recently, Yedidia et al. [7] demonstrated an intriguing connection between BBP and certain inference methods based in statistical physics. Essentially, they demonstrated that traditional BBP algorithms can be shown to arise from approximations of the extrema of the Bethe and Kikuchi free energies. Next, Yuille [8] derived new double-loop algorithms which are guaranteed to minimize the Bethe and Kikuchi energy functions while continuing to have close ties to the original BBP algorithms. Yuille?s approach relies on a certain decomposition of the Bethe and Kikuchi free energies. In the present work, we begin with a new principle?pairwise mutual information minimization and marginal entropy maximization (MIME)?and derive a new energy function which is shown to be equivalent to the Bethe free energy. After demonstrating this connection, we derive a family of free energies closely related to the MIME principle which also shown to be equivalent, when constraint satisfaction is exact, to the Bethe free energy. For each member in this family of energy functions , we derive a new algorithm that is guaranteed to converge to a local minimum. Moreover, the resulting form of the algorithm is very simple despite the somewhat unwieldy nature of the algebraic development. Preliminary comparisons of the new algorithm with BBP were carried out on spin glass-like problems and indicate that the new algorithm is convergent when BBP is not. However, the effectiveness of the new algorithms remains to be seen. 2 Bethe free energy and the MIME principle In this section, we show that the Bethe free energy can be interpreted as pairwise mutual information minimization and marginal entropy maximization. The Bethe free energy for Bayesian belief propagation is written as FBethe ({pij , pi , ?ij , ?ij }) = P P P P pij (xi ,xj ) pi (xi ) ij:i>j xi ,xj pij (xi , xj ) log ?ij (xi ,xj ) ? i (ni ? 1) xi pi (xi ) log ?i (xi ) P P P + ij:i>j xj ?ij (xj )[ xi pij (xi , xj ) ? pj (xj )] P P P + ij:i>j xi ?ji (xi )[ xj pij (xi , xj ) ? pi (xi )] P P + ij:i>j ?ij ( xi ,xj pij (xi , xj ) ? 1) (1) def where ?ij (xi , xj ) = ?ij (xi , xj )?i (xi )?j (xj ) and ni is the number of neighbors of node i. Link functions ?ij > 0 are available relational data between nodes i and j. The P singleton function ?i is also available at each node i. The double summation ij:i>j is carried out only over the nodes that are connected. The Lagrange parameters {?ij , ?ij } are needed in the Bethe free energy (1) to satisfy the following constraints relating the joint probabilities {pij } with the marginals {pi }: X X X pij (xi , xj ) = pj (xj ), pij (xi , xj ) = pi (xi ), and pij (xi , xj ) = 1. (2) xi xi ,xj xj The pairwise mutual information is defined as X pij (xi , xj ) M Iij = pij (xi , xj ) log p i (xi )pj (xj ) x ,x i (3) j The mutual information is minimized when the joint probability pij (xi , xj ) = pi (xi )pj (xj ) or equivalently when nodes i and j are independent. When nodes i and j are connected via a non-separable link ?ij (xi , xj ) they will not be independent. We now state the MIME principle. Statement of the MIME principle: Maximize the marginal entropy and minimize the pairwise mutual information using the available marginal and pairwise link function expectations while satisfying the joint probability constraints. The pairwise MIME principle leads to the following free energy: FMIME ({pij , pi , ?ij , ?ij }) = P P P P pij (xi ,xj ) i xi pi (xi ) log pi (xi ) ij:i>j xi ,xj pij (xi , xj ) log pi (xi )pj (xj ) + P P P P ? ij:i>j xi ,xj pij (xi , xj ) log ?ij (xi , xj ) ? i xi pi (xi ) log ?i (xi ) P P P + ij:i>j xj ?ij (xj )[ xi pij (xi , xj ) ? pj (xj )] P P P + ij:i>j xi ?ji (xi )[ xj pij (xi , xj ) ? pi (xi )] P P + ij:i>j ?ij ( xi ,xj pij (xi , xj ) ? 1). (4) In the above free energy, we minimize the pairwise mutual information and maximize the marginal entropies. The singleton and pairwise link functions are additional information which do not allow the system to reach its ?natural? equilibrium?a uniform i.i.d. distribution on the nodes. The Lagrange parameters enforce the constraints between the pairwise and marginal probabilities. These constraints are the same as in the Bethe free energy (1). Note that the Lagrange parameter terms vanish if the constraints in (2) are exactly satisfied. This is an important point when considering equivalences between different energy functions. Lemma 1 Provided the constraints in (2) are exactly satisfied, the MIME free energy in (4) is equivalent to the Bethe free energy in (1). Proof: Using the fact that constraint satisfaction is exact and using the identity pij (xi , xj ) = pji (xj , xi ), we may write X X X X pij (xi , xj ) log pi (xi ) pij (xi , xj ) log pi (xi )pj (xj ) = ? ? ij:i>j xi ,xj ij:i6=j xi ,xj =? X ni X X ij:i>j xi ,xj pij (xi , xj ) log ?i (xi )?j (xj ) = X i pi (xi ) log pi (xi ), xi i and X ni X pi (xi ) log ?i (xi ). (5) xi We have shown that a marginal entropy term emerges from the mutual information term in (4) when constraint satisfaction is exact. Collecting the marginal entropy terms together and rearranging the MIME free energy in (4), we get the Bethe free energy in (1). 3 A family of decompositions of the Bethe free energy Recall that the Bethe free energy and the energy function resulting from application of the MIME principle were shown to be equivalent. However, the MIME energy function is merely one particular decomposition of the Bethe free energy. As Yuille mentions [8], many decompositions are possible. The main motivation for considering alternative decompositions is for algorithmic reasons. We believe that certain decompositions may be more effective than others. This belief is based on our previous experience with closely related deterministic annealing algorithms [3, 2]. In this section, we derive a family of free energies that are equivalent to the Bethe free energy provided constraint satisfaction is exact. The family of free energies is inspired by and closely related to the MIME free energy in (4). Lemma 2 The following family of energy functions indexed by the free parameters ? > 0 and {?i } is equivalent to the original Bethe free energy (1) provided the constraints in (2) are exactly satisfied and the parameters q and r are set to {q i = (1 ? ?)ni } and {ri = 1 ? ni ?i } respectively. Fequiv ({pij , pi , ?ij , ?ij }) = P P P ij:i>j xi ,xj pij (xi , xj ) log [ xj P P P pij (xiP ,xj ) pij (xi ,xj )]? [ pij (xi ,xj )]? xi P + i xi pi (xi ) log pi (xi ) ? i qi xi pi (xi ) log pi (xi ) P P ? ? ij:i>j xi ,xj pij (xi , xj ) log ?ij (xi , xj )?i?i (xi )?j j (xj ) P P ? i ri xi pi (xi ) log ?i (xi ) P P P + ij:i>j xj ?ij (xj )[ xi pij (xi , xj ) ? pj (xj )] P P P + ij:i>j xi ?ji (xi )[ xj pij (xi , xj ) ? pi (xi )] P P + ij:i>j ?ij ( xi ,xj pij (xi , xj ) ? 1). (6) In (6), the first term is no longer the pairwise mutual information as in (4). And unlike (4), pi (xi ) no longer appears in the pairwise mutual information-like term. P P Proof: We selectively substitute xj pij (xi , xj ) = xi pij (xi , xj ) = pj (xj ) and pi (xi ) to show the equivalence. First X X X X X X pij (xi , xj )]? [ pij (xi , xj ) log[ ij:i>j xi ,xj xj X X pij (xi , xj )]? = ? ? xj i ? pij (xi , xj ) log ?i i (xi )?j j (xj ) = ij:i>j xi ,xj pi (xi ) log pi (xi ), ni xi X n i ?i i X pi (xi ) log ?i (xi ). (7) xj Substituting the identities in (7) into (6), we see that the free energies are algebraically equivalent. 4 A family of algorithms for belief propagation We now derive descent algorithms for the family of energy functions in (6). All the algorithms are guaranteed to converge to a local minimum of (6) under mild assumptions regarding the number of fixed points. For each member in the family of energy functions, there is a corresponding descent algorithm. Since the form of the free energy in (6) is complex and precludes easy minimization, we use algebraic (Legendre) transformations [1] to simplify the optimization. X X pij (xi , xj ) = pij (xi , xj ) log ? xj xj min?ji (xi ) ? ? X P xj pij (xi , xj ) log xi pij (xi , xj ) log ?ji (xi ) + ?ji (xi ) ? X pij (xi , xj ) = P xj pij (xi , xj ) xi min?ij (xj ) ? P xi pij (xi , xj ) log ?ij (xj ) + ?ij (xj ) ? P xi pij (xi , xj ) ?pi (xi ) log pi (xi ) = min ?pi (xi ) log ?i (xi ) + ?i (xi ) ? pi (xi ). ?i (xi ) (8) We now apply the above algebraic transforms. The new free energy is (after some algebraic manipulations) X X pij (xi , xj ) Fequiv ({pij , pi , ?ij , ?i , ?ij , ?ij }) = pij (xi , xj ) log ij:i>j xi ,xj ? (x )? ? (x ) ?ji i j ij +? XX ij:i6=j ? X X ?ij (xj ) + xi XX i pi (xi ) log xi ? ? pij (xi , xj ) log ?ij (xi , xj )?i i (xi )?j j (xj ) ? XX ij:i>j ?ij (xj )[ xj + q ?i i (xi ) X XX pij (xi , xj ) ? pj (xj )] + xi ij:i>j X X ri xi qi ?i (xi ) xi pi (xi ) log ?i (xi ) xi ?ji (xi )[ + X X i i ij:i>j xi ,xj + pi (xi ) X pij (xi , xj ) ? pi (xi )] xj X ij:i>j ?ij ( X pij (xi , xj ) ? 1). (9) xi ,xj We continue to keep the parameters {qi } and {ri } in (9). However, from Lemma 2, we know that the equivalence of (9) to the Bethe free energy is predicated upon appropriate setting of these parameters. In the rest of the paper, we continue to use q and r for the sake of notational simplicity. Despite the introduction of new variables via Legendre transforms, the optimization problem in (9) is still a minimization problem over all the variables. The algebraically transformed energy function in (9) is separately convex w.r.t. {p ij , pi } and w.r.t. {?ij , ?i } provided ? ? [0, 1]. Since the overall energy function is not convex w.r.t. all the variables, we pursue an alternating algorithm strategy similar to the double loop algorithm in Yuille [8]. The basic idea is to separately minimize w.r.t. the variables {?ij , ?i } and the variables {pij , pi }. The linear constraints in (2) are enforced when minimizing w.r.t the latter and do not affect the convergence properties of the algorithm since the energy function w.r.t. {pij , pi } is convex . We evaluate the fixpoints of {?ij , ?i }. Note that (9) is convex w.r.t. {?ij , ?i }. X X pij (xi , xj ), and ?i (xi ) = pi (xi ). (10) pij (xi , xj ), ?ji (xi ) = ?ij (xj ) = xj xi The fixpoints of {pij , pi } are evaluated next. Note that (9) is convex w.r.t. {pij , pi }. ? ? ? = ?ji (xi )?ij (xj )?ij (xi , xj )?i?i (xi )?j j (xj )e??ij (xj )??ji (xi )??ij ?1 P pi (xi ) = ?qi i (xi )?iri (xi )e k ?ki (xi )?1 . (11) pij (xi , xj ) The constraint satisfaction equations from (2) can be rewritten as X pij (xi , xj ) = pi (xi ) ? xj e 2?ji (xi ) = P xj ? ? ? ? ?ji (xi )?ij (xj )?ij (xi ,xj )?i i (xi )?j j (xj )e??ij (xj )??ij ?1 q r ?i i (xi )?i i (xi )e P k6=j ?ki (xi )?1 (12) Similar relations can be obtained for the other constraints in (2). Consider a Lagrange parameter update sequence where the Lagrange parameter currently being updated is tagged as ?new? with the rest designated as ?old.? We can then rewrite the Lagrange parameter updates using ?old? and ?new? values. Please note that each Lagrange parameter update corresponds to one of the constraints in (2). It can be shown that the iterative update of the Lagrange parameters is guaranteed to converge to the unique solution of (2) [8]. While rewriting (12), we multiply the old left and right sides with e?2?ji (xi ) . new e2?ji P (xi )?2?old ji (xi ) xj = ? ??old (xj )??old (xi )?? old ?1 ? ? ? ji ij ?ji (xi )?ij (xj )?ij (xi ,xj )?i i (xi )?j j (xj )e ij q r ?i i (xi )?i i (xi )e P k ?old (xi )?1 ki . (13) Using (11), we relate each Lagrange parameter update with an update of p ij (xi , xj ) and pi (xi ). We again invoke the ?old? and ?new? designations, this time on the probabilities. From (11), (12) and (13), we write the joint probability update s pnew old pold (xi ) ij (xi , xj ) ??new (x )+? (x ) (14) = e ji i ji i = P i old old pij (xi , xj ) xj pij (xi , xj ) and for the marginal probability update new old pnew (xi ) i = e?ji (xi )??ji (xi ) = pold (x ) i i sP xj pold ij (xi , xj ) pold i (xi ) (15) . From (14) and (15), the update equations for the probabilities are s s X pold (xi ) new old pij (xi , xj ) = pij (xi , xj ) P i old , pnew (x ) = pold pold i i i (xi ) ij (xi , xj ) xj pij (xi , xj ) x j (16) With the probability updates in place, we may write down new algorithms minimizing the family of Bethe equivalent free energies using only probability updates. The update equations (16) can be seen to satisfy the first constraint in (2). Similar update equations can be derived for the other constraints in (2). For each Lagrange parameter update, an equivalent, simultaneous probability (joint and marginal) update can be derived similar to (16). The overall family of algorithms can be summarized as shown in the pseudocode. Despite the unwieldy algebraic development preceding it, the algorithm is very simple and straightforward. Set free parameters ? ? [0, 1] and {?i }. Initialize {pij , pi }. Set {qi = (1 ? ?)ni } and {ri = 1 ? ni ?i }. Begin A: Outer Loop P ?ij (xj ) ? xi pij (xi , xj ) P ?ji (xi ) ? xj pij (xi , xj ) ?i (xi ) ? pi (xi ) ? ? ? pij (xi , xj ) ? ?ji (xi )?ij (xj )?ij (xi , xj )?i?i (xi )?j j (xj ) pi (xi ) ? ?qi i (xi )?iri (xi ) P P Begin B: Inner Loop: Do B until N1 ij:i>j [( xj pij (xi , xj ) ? P pi (xi ))2 + ( xi pij (xi , xj ) ? pj (xj ))2 ] < cthr Simultaneously update pij (xi , xj ) and pi (xi ) below. r pij (xi , xj ) ? pij (xi , xj ) P pi (xi ) x pij (xi ,xj ) j q P pi (xi ) ? pi (xi ) xj pij (xi , xj ) Simultaneously update pij (xi , xj ) and pj (xj ) below. r p (x ) pij (xi , xj ) ? pij (xi , xj ) P j j x pij (xi ,xj ) i q P pj (xj ) ? pj (xj ) xi pij (xi , xj ) Normalize pij (xi , xj ). p (x ,x ) pij (xi , xj ) ? P ij i j xi ,xj pij (xi ,xj ) End B End A In the above family of algorithms, the MIME algorithm corresponds to free parameter settings ? = 1 and ?i = 0 which in turn lead to parameter settings qi = 0 and ri = 1. The Yuille [8] double loop algorithm corresponds to the free parameter settings ? = 0 and ?i = 0 which in turn leads to parameter settings qi = ni and ri = 1. A crucial point is that the energy function for every valid parameter setting is equivalent to the Bethe free energy provided constraint satisfaction is exact. The inner loop constraint satisfaction threshold parameter cthr setting is very important in this regard. We are obviously not restricted to the MIME parameter settings. At this early stage of exploration of the inter-relationships between Bayesian belief propagation and inference methods based in statistical physics [7], it is premature to speculate regarding the ?best? parameter settings for ? and {?i }. Most likely, the effectiveness of the algorithms will vary depending on the problem setting which enters into the formulation via the link functions {?ij } and the singleton functions {?i }. 5 Results We implemented the family of algorithms in C++ and conducted tests on locally connected 50 node graphs and binary state variables. The ?i (xi ) and ?ij (xi , xj ) are of the form e?hi and e?hij where hi and hij are drawn from uniform distributions (in the interval [?1, 1]). Provided the constraint satisfaction theshold parameter cthr was set low enough, the algorithm (for ? = 1 and other parameter settings as described in Figure 1) exhibited monotonic convergence. Figure 2 shows the number of inner loop iterations corresponding to different settings of the constraint satisfaction threshold parameter. We also implemented the BBP algorithm and empirically observed that it often did not converge for these graphs. These results are quite preliminary and far more validation experiments are required. However, they provide a proof of concept for our approach. 6 Conclusion We began with the MIME principle and showed the equivalence of the MIMEbased free energy to the Bethe free energy assuming constraint satisfaction to be exact. Then, we derived new decompositions of the Bethe free energy inspired by the MIME principle, and driven by our belief that certain decompositions may be more effective than others. We then derived a convergent algorithm for each member in the family of MIME-based decompositions. It remains to be seen if the MIME-based algorithms are efficient for a reasonable class of problems. While the MIME-based algorithms derived here use closed-form solutions in the constraint satisfaction inner loop, it may turn out that the inner loop is better handled using preconditioned gradient-based descent algorithms. And it is important to explore the inter-relationships between the convergent MIME-based descent algorithms and other recent related approaches with interesting convergence properties [4, 5]. References [1] E. Mjolsness and C. Garrett. Algebraic transformations of objective functions. Neural Networks, 3:651?669, 1990. [2] A. Rangarajan. Self annealing and self annihilation: unifying deterministic annealing and relaxation labeling. Pattern Recognition, 33:635?649, 2000. [3] A. Rangarajan, S. Gold, and E. Mjolsness. A novel optimizing network architecture with applications. Neural Computation, 8(5):1041?1060, 1996. [4] Y. W. Teh and M. Welling. Passing and bouncing messages for generalized inference. Technical Report GCNU 2001-01, Gatsby Computational Neuroscience Unit, University College, London, 2001. [5] M. Wainwright, T. Jaakola, and A. Willsky. Tree-based reparameterization framework for approximate estimation of stochastic processes on graphs with cycles. Technical Report LIDS P-2510, MIT, Cambridge, MA, 2001. [6] Y. Weiss. Correctness of local probability propagation in graphical models with loops. Neural Computation, 12:1?41, 2000. [7] J. S. Yedidia, W. T. Freeman, and Y. Weiss. Bethe free energy, Kikuchi approximations and belief propagation algorithms. In Advances in Neural Information Processing Systems 13, Cambridge, MA, 2001. MIT Press. [8] A. L. Yuille. A double loop algorithm to minimize the Bethe and Kikuchi free energies. Neural Computation, 2001. (submitted). ?0.5 0 500 1000 MIME energy ?0.4 MIME energy ?0.4 MIME energy ?0.4 ?0.5 1500 0 500 1000 ?0.5 1500 0 500 1000 iteration iteration iteration (a) (b) (c) 1500 Figure 1: MIME energy versus outer loop iteration: 50 node, local topology, ? = 1. Constraint satisfaction threshold parameter cthr was set to (a) 10?8 (b) 10?4 (c) 10?2 7 20 2 18 1.9 6 1.8 12 10 8 5 total # of inner loop iterations 14 total # of inner loop iterations total # of inner loop iterations 16 4 3 6 1.7 1.6 1.5 1.4 1.3 4 1.2 2 2 0 1.1 0 500 1000 outer loop iteration index (a) 1500 1 0 500 1000 outer loop iteration index (b) 1500 1 0 500 1000 1500 outer loop iteration index (c) Figure 2: Inner loop iterations versus outer loop: 50 node, local topology, ? = 1. Constraint satisfaction threshold parameter cthr was set to (a) 10?8 (b) 10?4 (c) 10?2
2009 |@word mild:1 fixpoints:2 simulation:1 gainesville:1 decomposition:12 pold:7 mention:2 intriguing:1 written:1 update:17 smith:1 node:11 org:1 kettlewell:1 pairwise:12 inter:2 inspired:2 freeman:1 considering:2 begin:4 provided:6 moreover:1 xx:4 interpreted:1 pursue:1 extremum:2 transformation:2 every:1 collecting:1 tie:2 exactly:3 unit:1 engineering:1 local:7 despite:3 equivalence:4 jaakola:1 unique:1 get:1 close:2 fruitful:1 equivalent:10 demonstrated:3 deterministic:2 straightforward:1 iri:2 convex:5 simplicity:1 reparameterization:1 updated:1 exact:6 agreement:1 satisfying:1 recognition:1 observed:1 enters:1 connected:3 cycle:1 mjolsness:2 principled:1 fbethe:1 rewrite:1 yuille:10 upon:1 joint:5 effective:2 london:1 labeling:1 quite:1 precludes:1 obviously:1 sequence:1 loop:21 gold:1 normalize:1 convergence:3 double:6 rangarajan:3 kikuchi:5 derive:5 develop:1 depending:1 ij:87 implemented:2 indicate:1 closely:3 stochastic:1 exploration:1 preliminary:3 summation:1 annihilation:1 equilibrium:1 algorithmic:1 substituting:1 vary:1 early:1 estimation:1 currently:1 correctness:1 minimization:6 mit:2 always:1 derived:7 notational:1 glass:1 inference:4 xip:1 relation:1 transformed:1 overall:2 k6:1 development:3 initialize:1 mutual:11 marginal:12 construct:1 minimized:1 others:2 report:2 simplify:1 simultaneously:2 n1:1 message:1 multiply:1 experience:1 indexed:1 tree:1 continuing:1 old:15 theoretical:1 maximization:4 uniform:2 conducted:1 st:1 physic:3 invoke:1 together:1 again:1 satisfied:3 singleton:3 speculate:1 summarized:1 satisfy:2 mime:25 closed:1 minimize:5 spin:1 ni:10 correspond:1 yield:1 bayesian:5 submitted:1 simultaneous:1 reach:1 energy:59 e2:1 gcnu:1 proof:3 recall:1 emerges:1 garrett:1 appears:1 wei:2 formulation:1 evaluated:1 predicated:1 stage:1 until:1 propagation:9 believe:1 concept:1 tagged:1 alternating:1 self:2 please:1 generalized:1 novel:1 recently:2 began:1 pseudocode:1 ji:24 empirically:1 he:1 interpretation:1 relating:1 marginals:1 cambridge:2 i6:2 longer:2 showed:1 recent:1 optimizing:1 driven:1 manipulation:1 certain:5 binary:1 continue:2 seen:3 minimum:4 additional:1 somewhat:1 preceding:1 converge:6 maximize:2 algebraically:2 alan:1 technical:2 dept:1 qi:8 basic:1 essentially:1 expectation:1 iteration:12 separately:2 annealing:3 interval:1 crucial:1 rest:2 unlike:1 exhibited:1 ufl:1 member:3 anand:2 effectiveness:2 easy:1 enough:1 xj:179 affect:1 architecture:1 topology:2 inner:9 regarding:2 idea:1 handled:1 algebraic:6 passing:1 transforms:2 locally:1 neuroscience:1 theshold:1 write:3 pnew:3 demonstrating:1 threshold:4 drawn:1 pj:14 rewriting:1 graph:3 relaxation:1 merely:1 enforced:1 bouncing:1 place:1 family:16 reasonable:2 fl:1 def:1 ki:3 guaranteed:6 hi:2 convergent:3 constraint:25 ri:7 sake:1 min:3 separable:1 designated:1 legendre:2 lid:1 restricted:1 equation:4 remains:2 turn:3 needed:1 know:1 end:2 available:3 rewritten:1 yedidia:3 apply:1 enforce:1 appropriate:1 alternative:1 pji:1 florida:1 substitute:1 original:3 graphical:3 unifying:1 objective:1 strategy:1 traditional:1 gradient:1 link:5 outer:6 reason:1 preconditioned:1 willsky:1 assuming:1 index:3 relationship:2 minimizing:2 equivalently:1 statement:1 relate:1 hij:2 ski:1 teh:1 descent:4 relational:1 required:1 connection:2 below:2 pattern:1 belief:10 wainwright:1 satisfaction:13 natural:1 eye:1 carried:2 designation:1 interesting:1 versus:2 validation:1 pij:86 principle:9 pi:56 free:46 side:1 allow:1 institute:1 neighbor:1 regard:1 valid:1 san:1 premature:1 far:1 welling:1 approximate:1 keep:1 francisco:1 xi:248 spectrum:1 iterative:1 bethe:29 nature:1 ca:1 rearranging:1 complex:1 sp:1 did:1 main:1 motivation:1 arise:1 fillmore:1 gatsby:1 iij:1 vanish:1 unwieldy:2 down:1 entropy:8 likely:1 explore:1 lagrange:10 monotonic:1 corresponds:3 relies:1 ma:2 identity:2 cise:1 bbp:7 lemma:3 called:1 total:3 selectively:1 college:1 latter:1 evaluate:1
1,107
201
364 Jain and Waibel Incremental Parsing by Modular Recurrent Connectionist Networks Ajay N. Jain Alex H. Waibel School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 ABSTRACT We present a novel, modular, recurrent connectionist network architecture which learns to robustly perform incremental parsing of complex sentences. From sequential input, one word at a time, our networks learn to do semantic role assignment, noun phrase attachment, and clause structure recognition for sentences with passive constructions and center embedded clauses. The networks make syntactic and semantic predictions at every point in time, and previous predictions are revised as expectations are affirmed or violated with the arrival of new information. Our networks induce their own "grammar rules" for dynamically transforming an input sequence of words into a syntactic/semantic interpretation. These networks generalize and display tolerance to input which has been corrupted in ways common in spoken language. 1 INTRODUCTION Previously, we have reported on experiments using connectionist models for a small parsing task using a new network formalism which extends back-propagation to better fit the needs of sequential symbolic domains such as parsing (Jain, 1989). We showed that connectionist networks could learn the complex dynamic behavior needed in parsing. The task included passive sentences which require dynamic incorporation of previously unseen right context information into partially built syntactic/semantic interpretations. The trained parsing network exhibited predictive behavior and was able to modify or confirm Incremental Parsing by Modular Recurrent Connectionist Networks IInterclause Units I I Clause RolfS Units Clause M Phra-s-e1-'1 .. . Phn-,-eJ-" Clause 1 Phra-s-e1""""11 .. . Phn-s-e-'1 r-I r-I I I II r-I Clause Structure Units r-I II IPhrase Level Gating U n l t s l t - - - - - - - 1 Word Level , Word Units I Figure 1: High-level Parsing Architecture. hypotheses as sentences were sequentially processed. It was also able to generalize well and tolerate iII-formed input In this paper, we describe work on extending our parsing architecture to grammatically complex sentences. 1 The paper is organized as follows. First, we briefly outline the network formalism and the general architecture. Second, the parsing task is defined and the procedure for constructing and training the parser is presented. Then the dynamic behavior of the parser is illustrated, and the performance is characterized. 2 NETWORK ARCHITECTURE We have developed an extension to back-propagation networks which is specifically designed to perform tasks in sequential domains requiring symbol manipulation (Jain, 1989). It is substantially different from other connectionist approaches to sequential problems (e.g. Elman, 1988; Jordan, 1986; Waibel et al., 1989). There are four major features of this formalism. One, units retain partial activation between updates. They can respond to repetitive weak stimuli as well as singular sharp stimuli. Two, units are responsive to both static activation values of other units and their dynamic changes. Three, well-behaved symbol buffers can be constructed using groups of units whose connections are gated by other units. Four. the formalism supports recurrent networks. The networks are able to learn complex time-varying behavior using a gradient descent procedure via error back-propagation. Figure 1 shows a high-level diagram of the general parsing architecture. It is organized into five hierarchical levels: Word, Phrase, Clause Structure, Clause Roles, and Inter1 Another presentation of this work appears in Jain and Waibel (1990). 365 366 Jain and Waibel clause. The description will proceed bottom up. A word is presented to the network by stimulating its associated word unit for a short time. This produces a pattern of activation across the feature units which represents the meaning of the word. The connections from the word units to the feature units which encode semantic and syntactic information about words are compiled into the network and are fixed. 2 The Phrase level uses the sequence of word representations from the Word level to build contiguous phrases. Connections from the Word level to the Phrase level are modulated by gating units which learn the required conditional assignment behavior. The Clause Structure level maps phrases into the constituent clauses of the input sentence. The Clause Roles level describes the roles and relationships of the phrases in each clause of the sentence. The final level, Interclause, represents the interrelationships among the clauses. The following section defines a parsing task and gives a detailed description of the construction and training of a parsing network which performs the task. 3 INCREMENTAL PARSING In parsing spoken language, it is desirable to process input one word at a time as words are produced by the speaker and to incrementally build an output representation. This allows tight bi-directional coupling of the parser to the underlying speech recognition system. In such a system, the parser processes information as soon as it is produced and provides predictive information to the recognition system based on a rich representation of the current context As mentioned earlier, our previous work applying connectionist architectures to a parsing task was promising. The experiment described below extends our previous work to grammatically complex sentences requiring a significant scale increase. 3.1 Parsing Task The domain for the experiment was sentences with up to three clauses including nontrivial center-embedding and passive constructions.3 Here are some example sentences: ? Fido dug up a bone near the tree in the garden. ? I know the man who John says Mary gave the book. ? The dog who ate the snake was given a bone. Given sequential input, one word at a time, the task is to incrementally build a representation of the input sentence which includes the following infonnation: phrase structure, clause structure, semantic role assignment, and interclause relationships. Figure 2 shows a representation of the desired parse of the last sentence in the list above. 2Connectionist networks have been used for lexical acquisition successfully (Miikkulainen and Dyer, 1989). However, in building large systems, it makes sense from an efficiency perspective to precompile as much lexical information as possible into a network. This is a pragmatic design choice in building large systems. 3The training set contained over 200 sentences. These are a subset of the sentences which form the example set of a parser based on a left associative grammar (Hausser, 1988). These sentences are grammatically interesting, but they do not reflect the statistical structure of common speech. Incremental Parsing by Modular Recurrent Connectionist Networks [Clause 1: [Clause 2: [The dog RECIP] [was given ACTION] [a bone PATIENT]] [who AGENT] [ate ACTION] [the snake PATIENT] (RELATIVE to Clause 1, Phrase 1)) Figure 2: Representation of an Example Sentence. 3.2 Constructing the Parser The architecture for the network follows that given in Figure 1. The following paragraphs describe the detailed network structure bottom up. The constraints on the numbers of objects and labels are fixed for a particular network. but the architecture itself is scalable. Wherever possible in the network construction. modularity and architectural constraints have been exploited to minimize training time and maximize generalization. A network was constructed from three separate recurrent subnetworks trained to perform a portion of the parsing task on the training sentences. The performance of the full network will be discussed in detail in the next section. The Phrase level contains three types of units: phrase block units. gating units. and hidden units. There are 10 phrase blocks. each being able to capture up to 4 words forming a phrase. The phrase blocks contain sets of units (called slots) whose target activation patterns correspond to word feature patterns of words in phrases. Each slot has an associated gating unit which learns to conditionally assign an activation pattern from the feature units of the Word level to the slot. The gating units have input connections from the hidden units. The hidden units have input connections from the feature units. gating units, and phrase block units. The direct recurrence between the gating and hidden units allows the gating units to learn to inhibit and compete with one another. The indirect recurrence arising from the connections between the phrase blocks and the hidden units provides the context of the current input word. The target activation values for each gating unit are dynamically calculated during training; each gating unit must learn to become active at the proper time in order to perform the phrasal parsing. Each phrase block with its associated gating and hidden units has its weights slaved to the other phrase blocks in the Phrase level. Thus. if a particular phrase construction is only present in one position in the training set. all of the phrase blocks still learn to parse the construction. The Clause Roles level also has shared weights among separate clause modules. This level is trained by simulating the sequential building and mapping of clauses to sets of units containing the phrase blocks for each clause (see Figure 1). There are two types of units in this level: labeling units and hidden units. The labeling units learn to label the phrases of the clauses with semantic roles and attach phrases to other (within-clause) phrases. For each clause. there is a set of units which assigns role labels (agent. patient. recipient. action) to phrases. There is also a set of units indicating phrasal modification. The hidden units are recurrently connected to the labeling units to provide context and competition as with the Phrase level; they also have input connections from the phrase blocks of a single clause. During training. the targets for the labeling units are set at the beginning of the input presentation and remain static. In order to minimize global error across the training set. the units must learn to become active or inactive as soon as 367 368 Jain and Waibel possible in the input. This forces the network to learn to be predictive. The Clause Structure and Interclause levels are trained simultaneously as a single module. There are three types of units at this level: mapping, labeling, and hidden units. The mapping units assign phrase blocks to clauses. The labeling units indicate relative clause and a subordinate clause relationships. The mapping and labeling units are recurrently connected to the hidden units which also have input connections from the phrase blocks of the Phrase level. The behavior of the Phrase level is simulated during training of this module. This module utilizes no weight sharing techniques. As with the Clause Roles level, the targets for the labeling and mapping units are set at the beginning of input presentation, thus inducing the same type of predictive behavior. 4 PARSING PERFORMANCE The separately trained submodules described above were assembled into a single network which performs the full parsing task. No additional training was needed to fine-tune the full parsing network despite significant differences between actual subnetwork performance and the simulated subnetwork performance used during training. The network successfully modeled the large diverse training set. This section discusses three aspects of the parsing network's performance: dynamic behavior of the integrated network, generalization, and tolerance to noisy input. 4.1 Dynamic Behavior The dynamic behavior of the network will be illustrated on the example sentence from Figure 2: "The dog who ate the snake was given a bone." This sentence was not in the training set. Due to space limitations, actual plots of network behavior will only be presented for a small portion of the network. Initially, all of the units in the network are at their resting values. The units of the phrase blocks all have low activation. The word unit corresponding to "the" is stimulated, causing its word feature representation to become active across the feature units of the Word level. The gating unit associated with the slot 1 of phrase block 1 becomes active, and the feature representation of "the" is assigned to the slot; the gate closes as the next word is presented. The remaining words of the sentence are processed similarly, resulting in the final Phrase level representation shown in Figure 2. While this is occurring, the higher levels of the network are processing the evolving Phrase level representation. The behavior of some of the mapping units of the Clause Structure Level is shown in Figure 3. Early in the presentation of the first word, the Clause Structure level hypothesizes that the first 4 phrase blocks will belong to the first clause-reflecting the dominance of single clause sentences in the training set. After "the" is assigned to the first phrase block, this hypothesis is revised. The network then believes that there is an embedded clause of 3 (possibly 4) phrases following the first phrase. This predictive behavior emerged spontaneously from the training procedure (a large majority of sentences in the training set beginning with a determiner had embedded clauses after the first phrase). The next two words ("dog who") confirm the network's expectation. The word "ate" allows the network to firmly decide on an embedded clause of 3 phrases within Incremental Parsing by Modular Recurrent Connectionist Networks ClllUse_1 The dog who ate the snake was gl\leO 1 1 1 1 1 1 1 1 1 1~1 1 1 1 1 1 ~1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 ClllUse_2 The dog who ate the snake was gi\leO -... -._-_............... -.... _............... -...................................... p h r a s ~ 111111111111 I111111 11111111111111,,"11111111111.. ..11111111111111111111111111111111111111 11111111 .. 11 .... 111,'11111111 .................... .,1111111111111111111111111111111111111111111111111111111111111111111111111111111111 p h r a s ~ ~ ~ .Illllllllll1l1llll11ll1l1l1lh .4 I ................... P i. . . ,ll m 111111111111111111111111111111111111 11111111111""'""'""'1111 ~I I I I I I I I I I I I I I j . . . . Idl l l l l l l l l l l l l l~~I I I I I I I I I I I I I I I I I I I I 11111111111111111111111111111111111111111111111111 111111111111111111111111111111111 .1111111111111111111111 II ........ .. 111"' ... 11 ...... . ......... . .. ......... """ .... ........ Figure 3: Example of Clause Structure Dynamic Behavior. the main clause. This is the correct clausal structure of the sentence and is confirmed by the remainder of the input. The Interclause level indicates the appropriate relative clause relationship during the initial hypothesis of the embedded clause. The Clause Roles level processes the individual clauses as they get mapped through the Clause Structure level. The labeling units for clause 1 initially hypothesize an Agent/Action/Patient role structure with some competition from a Rec/Act/Pat role structure (the Agent and Patient units' activation traces for clause I, phrase 1 are shown in Figure 4). This prediction occurs because active constructs outnumbered passive ones during training. The final decision about role structure is postponed until just after the embedded clause is presented. The verb phrase "was given" immediately causes the Rec/Act/Pat role structure to dominate. Also, the network indicates that a fourth phrase (e.g. "by Mary'') is expected to be the Agent. As with the first clause, an AgjAct/Pat role structure is predicted for clause 2; this time the prediction is borne out 4.2 Generalization One type of generalization is automatic. A detail of the word representation scheme was omitted from the previous discussion. The feature patterns have two parts: a syntactic/semantic part and an identification part. The representations of "John" and "Peter" differ only in their ID parts. Units in the network which learn do not have any input connections from the ID portions of the word units. Thus, when the network learns to 369 370 Jain and Waibel CUUiELPlfw.)El-[TIE_DXl The dog who ate the snake was given a bone I i ". . . . . ',", 1 1I1I1I1 I1Iml l ~ 1 ~1I1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1'1111111 ..... ""oIllm 111 . . . . . . . . . . . . . m Figure 4: Example of Clause Roles Dynamic Behavior. parse "John gave the bone to the dog:' it will know how to parse "Peter promised the mitt to the boy:' This type of generalization is extremely useful, both for addition of new words to the network and for processing many sentences not explicitly trained on. The network also generalizes to correctly process truly novel sentences-sentences which are distinct (ignoring ID features) from those in the training set. The weight sharing techniques at the Phrase and Clause Structure levels have an impact here. While being difficult to measure generalization quantitatively, some statements can be made about the types of novel sentences which can be correctly processed relative to the training sentences. Substitution of single words resulting in a meaningful sentence is tolerated almost without exception. Substitution of entire phrases by different phrases causes some errors in structural parsing on sentences which have few similar training exemplars. However, the network does quite well on sentences which can be formed from composition between familiar sentences (e.g. interchanging clauses). 4.3 Tolerance to Noise Several types of noise tolerance are interesting to analyze: ungrammaticality, word deletions (especially poorly articulated short function words), variance in word speed, interword silences, interjections, word/phrase repetitions, etc. The effects of noise were simulated by testing the parsing network on training sentences which had been corrupted in the ways listed above. Note that the parser was trained only on well-formed sentences. Sentences in which verbs were made ungrammatical were processed without difficulty (e.g. "We am happy."). Sentences in which verb phrases were badly corrupted produced reasonable interpretations. For example, the sentence "Peter was gave a bone to Fido:' received an AgJAct/Pat/Rec role structure as if "was gave" was supposed to be either "gave" or "has given". Interpretation of corrupted verb phrases was context dependent. Single clause sentences in which determiners were randomly deleted to simulate speech recognition errors were processed correctly 8S percent of the time. Multiple clause sentences degraded in a similar manner produced more parsing errors. There were fewer examples of multi-clause sentence types, and this hurt performance. Deletion of function words such as prepositions beginning prepositional phrases produced few errors, but deletions of critical function words such as "to" in infinitive constructions introducing subordinate clauses caused serious problems. Incremental Parsing by Modular Recurrent Connectionist Networks The network was somewhat sensitive to variations in word presentation speed (it was trained on a constant speed), but tolerated inter-word silences. Interjections of "ahhn and partial phrase repetitions were also tested. The network did not perform as well on these sentences as other networks trained for less complex parsing tasks. One possibility is that the weight sharing is preventing the formation of strong attractors for the training sentences. There appears to be a tradeoff between generalization and noise tolerance. 5 CONCLUSION We have presented a novel connectionist network architecture and its application to a non-trivial parsing task. A hierarchical, modular, recurrent connectionist network was constructed which successfully learned to parse grammatically complex sentences. The parser exhibited predictive behavior and was able to dynamically revise hypotheses. Techniques for maximizing generalization were also discussed. Network performance on novel sentences was impressive. Results of testing the parser's sensitivity to several types of noise were somewhat mixed, but the parser performed well on ungrammatical sentences and sentences with non-critical function word deletions. Acknowledgments This research was funded by grants from ATR Interpreting Telephony Research Laboratories and the National Science Foundation under grant number EET-87 16324. We thank Dave Touretzky for helpful comments and discussions. References J. L. Elman. (1988) Finding Structure in Time. Tech. Rep. 8801, Center for Research in Language, University of California, San Diego. R. Hausser. (1988) Computation of Language. Springer-Verlag. A. N. Jain. (1989) A Connectionist Architecturefor Sequential Symbolic Domains. Tech. Rep. CMU-CS-89-187, School of Computer Science, Carnegie Mellon University. A. N. Jain and A. H. Waibel. (1990) Robust connectionist parsing of spoken language. In Proceedings of the 1990 IEEE International Conference on Acoustics. Speech. and Signal Processing. M. I. Jordan. (1986) Serial Order: A Parallel Distributed Processing Approach. Tech. Rep. 8604, Institute for Cognitive Science, University of California, San Diego. R. Miikkulainen and M. O. Dyer. (1989) Encoding input/output representations in connectionist cognitive systems. In D. Touretzky. G. Hinton. and T. Sejnowski (eds.) , Proceedings of the 1988 Connectionist Models Summer School, pp. 347356. Morgan Kaufmann Publishers. A. Waibel, T. Hanazawa, O. Hinton, K. Shikano, and K. Lang. (1989) Phoneme recognition using time-delay neural networks. IEEE Transactions on Acoustics. Speech. and Signal Processing 37(3):328-339. 371
201 |@word briefly:1 idl:1 initial:1 substitution:2 contains:1 current:2 activation:8 lang:1 must:2 parsing:32 john:3 hypothesize:1 designed:1 plot:1 update:1 fewer:1 beginning:4 short:2 provides:2 five:1 constructed:3 direct:1 become:3 paragraph:1 manner:1 inter:1 expected:1 behavior:16 elman:2 multi:1 actual:2 affirmed:1 becomes:1 underlying:1 substantially:1 developed:1 spoken:3 finding:1 every:1 act:2 unit:61 grant:2 modify:1 despite:1 encoding:1 id:3 dynamically:3 bi:1 acknowledgment:1 spontaneously:1 testing:2 block:16 procedure:3 evolving:1 word:42 induce:1 symbolic:2 get:1 close:1 context:5 applying:1 map:1 center:3 lexical:2 maximizing:1 assigns:1 immediately:1 rule:1 dominate:1 embedding:1 variation:1 hurt:1 phrasal:2 construction:7 target:4 parser:10 diego:2 us:1 hypothesis:4 pa:1 recognition:5 rec:3 bottom:2 role:17 module:4 capture:1 connected:2 inhibit:1 mentioned:1 transforming:1 recip:1 dynamic:9 trained:9 tight:1 predictive:6 efficiency:1 indirect:1 leo:2 articulated:1 jain:10 distinct:1 describe:2 sejnowski:1 labeling:9 formation:1 whose:2 modular:7 emerged:1 quite:1 say:1 grammar:2 gi:1 unseen:1 syntactic:5 itself:1 noisy:1 final:3 associative:1 hanazawa:1 sequence:2 remainder:1 causing:1 fido:2 poorly:1 supposed:1 description:2 inducing:1 competition:2 constituent:1 extending:1 produce:1 incremental:7 object:1 coupling:1 recurrent:9 exemplar:1 interclause:4 school:3 received:1 strong:1 predicted:1 c:1 indicate:1 differ:1 correct:1 subordinate:2 require:1 assign:2 generalization:8 extension:1 mapping:6 major:1 early:1 omitted:1 determiner:2 label:3 infonnation:1 sensitive:1 repetition:2 successfully:3 ej:1 varying:1 encode:1 indicates:2 tech:3 sense:1 am:1 helpful:1 dependent:1 el:1 snake:6 integrated:1 entire:1 initially:2 hidden:10 i1:1 among:2 noun:1 construct:1 represents:2 interchanging:1 connectionist:17 stimulus:2 quantitatively:1 serious:1 few:2 randomly:1 simultaneously:1 national:1 individual:1 familiar:1 attractor:1 possibility:1 truly:1 partial:2 tree:1 desired:1 formalism:4 earlier:1 contiguous:1 assignment:3 phrase:54 introducing:1 subset:1 delay:1 reported:1 corrupted:4 tolerated:2 international:1 sensitivity:1 retain:1 reflect:1 containing:1 possibly:1 borne:1 interjection:2 book:1 cognitive:2 includes:1 hypothesizes:1 explicitly:1 caused:1 performed:1 bone:7 analyze:1 portion:3 parallel:1 dug:1 minimize:2 formed:3 degraded:1 variance:1 who:8 kaufmann:1 phoneme:1 correspond:1 directional:1 generalize:2 weak:1 identification:1 produced:5 confirmed:1 dave:1 touretzky:2 sharing:3 ed:1 acquisition:1 ungrammaticality:1 pp:1 associated:4 static:2 revise:1 organized:2 back:3 reflecting:1 appears:2 tolerate:1 higher:1 just:1 until:1 parse:5 propagation:3 incrementally:2 defines:1 behaved:1 mary:2 building:3 effect:1 requiring:2 contain:1 assigned:2 laboratory:1 semantic:8 illustrated:2 conditionally:1 ll:1 during:6 recurrence:2 speaker:1 outline:1 performs:2 interrelationship:1 interpreting:1 passive:4 percent:1 meaning:1 novel:5 common:2 clause:58 discussed:2 interpretation:4 belong:1 resting:1 mellon:2 significant:2 composition:1 automatic:1 similarly:1 language:5 had:2 funded:1 impressive:1 compiled:1 etc:1 own:1 showed:1 perspective:1 manipulation:1 buffer:1 verlag:1 rep:3 postponed:1 exploited:1 infinitive:1 morgan:1 additional:1 somewhat:2 maximize:1 signal:2 ii:3 full:3 desirable:1 multiple:1 characterized:1 serial:1 e1:2 impact:1 prediction:4 scalable:1 ajay:1 expectation:2 patient:5 cmu:1 repetitive:1 addition:1 separately:1 fine:1 diagram:1 singular:1 publisher:1 exhibited:2 comment:1 grammatically:4 jordan:2 structural:1 near:1 iii:1 submodules:1 fit:1 gave:5 architecture:10 tradeoff:1 inactive:1 peter:3 speech:5 proceed:1 cause:2 action:4 useful:1 detailed:2 listed:1 tune:1 processed:5 arising:1 correctly:3 clausal:1 diverse:1 carnegie:2 group:1 dominance:1 four:2 promised:1 deleted:1 compete:1 fourth:1 respond:1 extends:2 almost:1 reasonable:1 decide:1 architectural:1 utilizes:1 decision:1 summer:1 display:1 badly:1 nontrivial:1 incorporation:1 constraint:2 alex:1 aspect:1 speed:3 simulate:1 extremely:1 waibel:9 across:3 describes:1 ate:7 remain:1 prepositional:1 wherever:1 modification:1 previously:2 discus:1 needed:2 know:2 dyer:2 subnetworks:1 generalizes:1 hierarchical:2 appropriate:1 simulating:1 robustly:1 responsive:1 gate:1 recipient:1 remaining:1 build:3 especially:1 occurs:1 subnetwork:2 gradient:1 separate:2 mapped:1 simulated:3 atr:1 majority:1 thank:1 trivial:1 modeled:1 relationship:4 happy:1 difficult:1 statement:1 boy:1 trace:1 design:1 proper:1 perform:5 gated:1 revised:2 descent:1 pat:4 hinton:2 sharp:1 verb:4 dog:8 required:1 sentence:46 connection:9 california:2 hausser:2 learned:1 deletion:4 acoustic:2 assembled:1 able:5 below:1 pattern:5 rolf:1 built:1 including:1 garden:1 belief:1 critical:2 difficulty:1 force:1 attach:1 scheme:1 firmly:1 attachment:1 relative:4 embedded:6 mixed:1 interesting:2 limitation:1 telephony:1 foundation:1 agent:5 i111111:1 preposition:1 gl:1 last:1 soon:2 silence:2 institute:1 tolerance:5 distributed:1 calculated:1 phn:2 rich:1 preventing:1 made:2 san:2 miikkulainen:2 transaction:1 eet:1 confirm:2 global:1 sequentially:1 active:5 pittsburgh:1 shikano:1 modularity:1 stimulated:1 promising:1 learn:11 robust:1 ignoring:1 ungrammatical:2 complex:7 constructing:2 domain:4 did:1 main:1 noise:5 arrival:1 position:1 learns:3 gating:12 symbol:2 list:1 recurrently:2 sequential:7 occurring:1 forming:1 contained:1 partially:1 springer:1 slaved:1 stimulating:1 conditional:1 slot:5 presentation:5 shared:1 man:1 change:1 included:1 specifically:1 called:1 meaningful:1 indicating:1 pragmatic:1 exception:1 support:1 modulated:1 violated:1 tested:1
1,108
2,010
Information Geometrical Framework for Analyzing Belief Propagation Decoder Shiro Ikeda Kyushu Inst. of Tech., & PRESTO, JST Wakamatsu, Kitakyushu, Fukuoka, 808-0196 Japan shiro@brain.kyutech.ac.jp Toshiyuki Tanaka Tokyo Metropolitan Univ. Hachioji, Tokyo, 192-0397 Japan tanaka@eei.metro-u.ac.jp Shun-ichi Amari RIKEN BSI Wako, Saitama, 351-0198 Japan amari@brain.riken.go.jp Abstract The mystery of belief propagation (BP) decoder, especially of the turbo decoding, is studied from information geometrical viewpoint. The loopy belief network (BN) of turbo codes makes it difficult to obtain the true ?belief? by BP, and the characteristics of the algorithm and its equilibrium are not clearly understood. Our study gives an intuitive understanding of the mechanism, and a new framework for the analysis. Based on the framework, we reveal basic properties of the turbo decoding. 1 Introduction Since the proposal of turbo codes[2], they have been attracting a lot of interests because of their high performance of error correction. Although the thorough experimental results strongly support the potential of this iterative decoding method, the mathematical background is not sufficiently understood. McEliece et al.[5] have shown its relation to the Pearl?s BP, but the BN for the turbo decoding is loopy, and the BP solution gives only an approximation. The problem of the turbo decoding is a specific example of a general problem of marginalizing an exponential family distribution. The distribution includes higher order correlations, and its direct marginalization is intractable. But the partial model with a part of the correlations, can be marginalized with BP algorithm exactly, since it does not have any loop. By collecting and exchanging the BP results of the partial models, the true ?belief? is approximated. This structure is common among various iterative methods, such as Gallager codes, Beth?e approximation in statistical physics[4], and BP for loopy BN. We investigate the problem from information geometrical viewpoint[1]. It gives a new framework for analyzing these iterative methods, and shows an intuitive understanding of them. Also it reveals a lot of basic properties, such as characteristics of the equilibrium, the condition of stability, the cost function related to the decoder, and the decoding error. In this paper, we focus on the turbo decoding, because its structure is simple, but the framework is general, and the main results can be generalized. 2 Information Geometrical Framework 2.1 Marginalization, MPM Decoding, and Belief    which is defined as follows   !"$#%&!'("  )'* +',"-./$ (1) # "      / 1 2 " 3    is the linear function of 0 , and each is the higher order correlations where, of 0  1 2 . The problem of turbo codes and similar iterative methods are to marginalize 1=<>   1  this &    6 ;  9 7 8 :  distribution. Let 4 denote the operator of marginalization as, 45 . The marginalization is equivalent to take the expectation of as ?@6 ;798 A BC  & D?E&F   /F   HG In the case of MPM (maximization of the posterior marginals) decoding,  1JI 0LKM $' M 2 and the sign of each FN1 is the decoding result. In the belief network, 1 I 0PO M 2 and FN1 is the belief. In these iterative methods, the marginalization of eq.(1) is not tractable, but the marginalization of the following distribution is tractable. Q &SR/T* %U" # )'," Q &>'VT. KW Q XT/ ZY M  \[] ^T I`_  G (2) Each  Q &SRT includes only one of the 0 " 3 &$2 in eq.(1), and additional parameter T is used to adjust linear part of . The iterative methods are exchanging information through T for each aQ , and finally approximate 45  & . Let us consider a distribution of 2.2 The Case of Turbo Decoding bcNd9e%fghNi\f jNk\d | | l | m noXprqsXtvu w x | | m noXprqsXtNy {w z | l+} } ~?L???X?? | l+? } ? } ? | w x w\z w x w{z bcNd9e%f??k\i\f+jNk\d l+? } ? w$x || ? |? ? sXoXpr? qsX?tLu z ?? x sXoXprqsXt y l+? } ? wz Figure 1: Turbo codes ? ,??v\  /?v/? ?? ?&: ? ? 1 + 1  ? ? ?) ?%U?N ? ? ? I &S ?  ? ?  0LM  ?V2  S ? ?? ? ? ?? ?  ? 1 ?v? 9?N ? ? ? ? I 0vKM $' M 2 ?? M+? ? ? ? ? + ?? ?  . The ultimate goal of the turbo decoding is the MPM decoding of based on  ?? S Since the channel is memoryless, the following relation holds   S ? ?? ? ? ?? ? ? * %?? ? ?'??@? ? ? ? ?'??@?? ? ??SK ??' ?N? r???> ??? O ?  M  M?K??{? ??? ?> ???&?> 6 ?798 ? ? ? ??'(?N?? G ? In the case of turbo codes, is the information bits, from which the turbo encoder generates two sets of parity bits, , and , (Fig.1). Each parity bit is expressed as the form , where the product is taken over a subset of . The codeword is then transmitted over a noisy channel, which we assume BSC (binary symmetric channel) with flipping probability . The receiver observes , . 0LKM {' M 2 By assuming the uniform prior on , the posterior distribution is given as follows  B  S ? ? ? + ?? ? ?    %?&?  ? ?'?? ?   '?? ?  ? ? ? ? ? ? (3)  +S ? ?? ? + ?? ? ?   ?   U" # &!'(" +!'(" ? / G ? , " Q (?? ? ? Q ? Q Y M ?  . Here  is the normalizing factor, and "$#v&(?? S ^ [  ? . When ? is large, marginalization of to eq.(1), where ??rS ? ? ? ? ??? ? is equivalent Equation(3) is intractable since it needs summation over ? terms. Turbo codes utilize two decoders which solve the MPM decoding of aQ SRT?Y M ?  in eq.(2). The ? ? ? Q ?  and the prior of which has the form of distribution is derived from   ?  SRT   !XT K ??XT G  &SR/T is a factorizable distribution. The marginalization of   ? ? ? ? Q ?  is feasible since T its BN is loop free. The parameter serves as the window of exchanging the information between the two decoders. The MPM decoding is approximated by updating T iteratively  &?? ? ? ? ?  ? ? ?  in ?turbo? like way. 2.3 Information Geometrical View of MPM Decoding Let us consider the family of all the probability distributions over . We denote it by  , which is defined as         &  &S? O  I 0LKM $' M 2  A B   M G We consider an ? ?flat submanifold  # in  . This is the submanifold of  # SR  defined as     G  #   # &SR ? ?" # !' ` KW #  )? E    I _ (4) ? , every distribution of  # can be rewritten as follows Since " # &?*?? ' KW #    ???]'?   KW #  / G  # &SR ? U" # ! It shows that every distribution of  # is decomposable, or factorizable. From the information geometry[1], we have the following theorem of  ?projection. Theorem 1. Let  be an ? ?flat submanifold in  , and let  & I  . The point in  that minimizes the KL-divergence from  & to  , is denoted by, 4?5 &? ? %B $'& ?  $R  &-,U and is called the  ?projection of  & !#" to  )(+* . The  ?projection is unique. # It is easy to show that the marginalization corresponds to the  ?projection to  [7]. Since MPM decoding and marginalization is equivalent, MPM decoding is also equivalent to the  ?projection to  . # 2.4 Information Geometry of Turbo Decoding 5 & Let ./  denote the parameters in  ?5  ? of the  ?projected distribution, #& 576 ', 4 123  (8* .0 # ? The turbo decoding process is written as follows, $R  &SR  G T   O for   O , and   M .   by 2. Project  ? &SR/T  onto  # as  ./?5  ? &SR/T  , and calculate T ? T ?   .0P5  ? &SRT   K T  G       by 3. Project  &SR/T ?  onto  # as  .  ?5  &SRT ?  , and calculate T  T    ./?5   &SR/T ?   K T ?  G      .0P5  ? &SR/T    , go to step 2. 4. If .0P5   &SRT ?  the estimated parameter , the projection of ?? S ? turbo ? onto  # approximates ?  T V ' T ? , where the estimated distribution is The ?? ? ?? decoding  , as  # &SR  ?&" # &!'?T  P '?T ? KVW # XT  'VT ? / G (5)   T   eq.(5) is An intuitive understanding of the turbo decoding is as follows. In step 2, ? T   , and T ? is estimated byinprojecting replaced with " ? & . The distribution becomes  ? SR it onto  # . In step 3, XT   in eq.(5) is replaced with "  & , and T  is estimated by  ? T ?  . projection of  +SR We now define the submanifold corresponding to each decoder,       Q   Q SRT "$#v&!'(" Q !'?T P KVW Q XT ? T    I`_ Y? M ? G T is the coordinate system of  Q .  Q is also an ? ?flat submanifold.     ?? and  " ?  in general.  Q    # hold because " Q & includes cross terms of and " N&J 1. Let  The information geometrical view of the turbo decoding is schematically shown in Fig.2. 3 The Properties of Belief Propagation Decoder 3.1 Equilibrium  SRT    ? SRT ?   #LSR  M GG 45  +T &SR/'VT ? T H G 45  ? SRT  ?  # &SR  G  ? ?     as Let us define a manifold    &   I  ?A B  &? A B  # &SR   G   ?  From its definition, for any  & I    , the expectation of is the same, and its  ? projection to  # coincides with  # SR  . This is an  ?flat submanifold[1], and we call    an equimarginal submanifold. Since eq.(6) holds,  #LSR $  SRT ? $  ? SRT   I    is satisfied. an ? ?flat version of the submanifold as    , which connects  # SR  ,  , and  ? &SR/ T   in log-linear manner Let &SRus/ T define ?        # &SR    +&SR T ?    ? SR T     A ?  Q  M  G   Q < #  &?? ? ? ? ? ? ?? ?  is included in the    . It can be proved by taking Since  #  eq.(7)    holds,   KM , ? M . When the the turbo decoding converges, equilibrium solution defines three important dis , , and . They satisfy the following two conditions: tributions, (6) (7)                                 #% $'&)( +* #,  "!  "!              #   -  -/& . 2''3 457698 : ;=<>7?   - -/. ! !   0 # 1&! 1!        - -/. $ # ( +* ,  @  &! "&! Figure 2: Turbo decoding  A RUWvurwW x+wy pqxzwy t _ KLCFEHGJI BNQ R t UXWZY=s p ] _ BPO RpqUWrY9s t ] _ BDCFEHGJI BNM RTSVUXWZY\[^]`_ a+b=cVdfehgejik'l cJm=n k'o Figure 3:      and    # &SR   +SRT ?   &SR/T     ? ?  &?? ? ? ? ?  ? ? ?  ? ? ? + ?? ?  ,  &SR  is the true marginalization of  ?? S ? ? ? + ?? ?  . If    includes  ?? ?     ? ? ?  ? ? ?  . This fact means that necessarily include  ?? ? ??rS ? ? ? ? ??? ?  and does  are not necessarily equimarginal, which is the origin of the However,  # &SR not Theorem 2. When the turbo decoding procedure converges, the convergent probability , , and belong to equimarginal submanifold distributions    , while its ?flat version  includes these three distributions and also the posterior distribution (Fig.3).   decoding error. 3.2 Condition of Stability W # in eq.(4) and W Q in eq.(2) ?#L %6 E798 A B  #v&SR  L{ 4 W #L $ Z? Q XT%6 798 A B  Q &SRTL{}| W Q XT^Y? M ? G The expectation parameters are defined as follows with ? #  ?>+T ? ? ? T   G We give a sufficiently small perturbation ~ to T  and apply one turbo decoding step. The  ?projection from  ? &SRT ' ~  to  # gives, ? #  ? '  *? ? XT  ' ~   ?  ? #v  ?  ? ? XT   ~ G Here, ? #L  is the Fisher information matrix of  #v&SR  , and ? Q T is that of  Q &SRT , Y? M ?  . Note that ? #   is a diagonal matrix. The Fisher information matrix is defined as follows ? #  L  { 44"? W #  L  { 4 ? #  {? ? Q XT??  { |\| ? W Q XTL  { | ? Q XT ZY M ? G T ? in step 2 will be, T ?  T ? ? ' ?? #  $?  ? ? XT   K?? ? ? ~ G Here, ?  is an identity matrix of size ? . Following the same line for step 3, we derive the Equation (6) is rewritten as follows with these parameters,  theorem which coincides with the result of Richardson[6].  Theorem 3. Let When ? 1{? ? M 1 be the eigenvalues of the matrix  defined as  ? ? #% {?  ?  T ?  K ?  ?Z? ? #% $ ?  ? ? T   K?? ?? G  holds for all , the equilibrium point is stable. 3.3 Cost Function and Characteristics of Equilibrium  XT  /T ?  W L#   K  W  T ? !' W ? XT  / G *T '?T ? T  G G G /T - is the critical point of  . Theorem 4. The equilibrium state   T *  ? #   K ? ? XT?  , { |   *? #   K ?>?XT ?  . For the Proof. Direct calculation gives { |   ? L #       ?   T     ?  equilibrium, ? ?  holds, and the proof is completed. T Q  K T Q  is small, When X  T    T    ? #%  ?   {}|   G  T ?  K T ?  K ? #%  ?  {|  We give the cost function which plays an important role in turbo decoding.  Here, . This function is identical to the ?free energy? defined in [4].    {{ || ||  {{ || ||    ? # ? K # ?? ? # ? K # ? ? G   T '?T ?  T  K T ?   {  {   M   ? #   K  ??H' ? ?  X?? K ? ?  G  X?? K ? ?  K  ??'?? ?   { {     { {   This shows how the algorithm works, but it does not give the characteristics of the equilibrium point. The Hessian of is         And by transforming the variables as, 44 4 4  Most probably, 44 is positive definite but saddle at equilibrium. and , we have is always negative, and is generally  & SR ! as  SR !  ?" # !' ` ]'  KW  !/ ]   ?   W  !H ? ? A B   ?&"$#v&!' ' / a&%6 (798 &"  &$ \" ? &/  G   ,    ), and  Q &SR/T ? ? ?  ? ? ?  (  This distribution includes  #v&SR  ( ),  &?? S ( ,;T ,   Q ), where ?  M M / , @  M O / , and ?   O M / . The expectation parameter ?? ! is defined as, ?? ! L{ 4 W   ! A B  SR  ! G Let us consider    , where every distribution  SR  ! I    has the same expectation parameter, that is, ?S  !H ?S  holds. Here, we define, ?S H ?S +! a . From the Taylor expansion, we have, FL1 ! FL1 !' A { ?PFN1/ 9  ? ' A { Q FN1\ " Q ' M A { Q { $ FL1 " Q  $ Q ? ? Q# $ 'A { Q { ? F 1  "  Q   ? ' M A { 3 { % F 1  9   3   % ' ! &'(&)P!'  &1 *&+)  G (8) ? #Q ? 3 #% 3.4 Perturbation Analysis For the following discussion, we define a distribution   0  L F 1 r 2  !?EF 1  0 Y+  N2 ? F 1   J 1 a? 6  798  K 0 1? 2 1 &  SR ? ! a    Q Q  11   1 K 11 A 1 Q  Q K A { Q K A 33 3 Q { 3 { $ K A ?r? ? $ { ? F 1  " Q  $ ? 3 ? Q # $  Q (9) 1 1  1=1 L  { F    1 1 Q M?? , and where, . &  SR   $    +SR  I    holds,  T ? and     a   Let since T ? K  K , T and V     . Also when we put ? ,  ? K T ? holds. From eq.(9), we have the following result, 1 # 1=1 1 Q 1 1 { Q A 33 3 Q {%3 { Q A ?r? ? Q {N? FN1\  G Q (10) K  K K ? K 3 K ? Next, let    , and we consider  &SR ! % I    , where is the parameter which ? ? ?  ? ? ?  is not necessarily included in satisfies this equation. Since  SR!!     &?? S     , is generally not equal to . From eq.(9),  1 1# 1 1 A 1 Q 1 1 A { Q A 33 3 Q { 3 { Q A ?r? ? Q {N? FN1  G  K  K ?  K Q K1 ? 1 Q K 3 From the condition  T  'VT ? and eq.(10), we have the following approximation, 11 1  A { Q K A 33 3 Q { 3 { $ K A ?r? ? $ { ? FL1  G K ?  ? Q< $ 3 . After adding some The indexes    are for ,  are for , and  +   {    definitions, that is, , and , where  is the Fisher information matrix of which is a diagonal matrix, we substitute   with function of up to its 2nd order, and neglect the higher orders of . And we have,                            ? ? ?S  # ? ?? ) % ?? ! ' M A { Q K A 3 3 3 Q { 3 { $ K A ?r? ? $ { ? ?S  G ? 3 ? Q< $ This result gives the approximation accuracy of the BP decoding. Let the true belief be   , and we evaluate the difference between   and  on  . The result is summarized in the following theorem. Theorem 5. The true expectation of , which is   , is approximated as, ?   ??  Where      is the solution of the turbo decoding.   Equation (11) is related to the  ?embedded?curvature of   extended to general case where  [3, 8]. [  (11) (Fig.3). The result can be 4 Discussion We have shown a new framework for understanding and analyzing the belief propagation decoder. Since the BN of turbo codes is loopy, we don?t have enough theoretical results for BP algorithm, while a lot of experiments show that it works surprisingly well in such cases. The mystery of the BP decoders is summarized in 2 points, the approximation accuracy and the convergence property. Our results elucidate the mathematical background of the BP decoding algorithm. The information geometrical structure of the equilibrium is summarized in Theorem 2. It shows ?     the ?flat submanifold   plays an important role. Furthermore, Theorem 5 shows that the relation between   and the  ?flat submanifold   causes the decoding error, and the principal component of the error is the curvature of   . Since the curvature strongly depends on the codeword, we can control it by the encoder design. This shows a room for improvement of the ?near optimum error correcting code?[2].     For the convergent property, we have shown the energy function, which is known as Beth?e free energy[4, 9]. Unfortunately, the fixed point of the turbo decoding algorithm is generally a saddle of the function, which makes further analysis difficult. We have only shown a local stability condition, and the global property is one of our future works. This paper gives a first step to the information geometrical understanding of the belief propagation decoder. The main results are for the turbo decoding, but the mechanism is common with wider class, and the framework is valid for them. We believe further study in this direction will lead us to better understanding and improvements of these methods. Acknowledgments We thank Chiranjib Bhattacharyya who gave us the opportunity to face this problem. We are also grateful to Yoshiyuki Kabashima and Motohiko Isaka for useful discussions. References [1] S. Amari and H. Nagaoka. (2000) Methods of Information Geometry, volume 191 of Translations of Mathematical Monographs. American Mathematical Society. [2] C. Berrou and A. Glavieux. (1996) Near optimum error correcting coding and decoding: Turbo-codes. IEEE Transactions on Communications, 44(10):1261?1271. [3] S. Ikeda, T. Tanaka, and S. Amari. (2001) Information geometry of turbo codes and low-density parity-check codes. submitted to IEEE transaction on Information Theory. [4] Y. Kabashima and D. Saad. (2001) The TAP approach to intensive and extensive connectivity systems. In M. Opper and D. Saad, editors, Advanced Mean Field Methods ? Theory and Practice, chapter 6, pages 65?84. The MIT Press. [5] R. J. McEliece, D. J. C. MacKay, and J.-F. Cheng. (1998) Turbo decoding as an instance of Pearl?s ?belief propagation? algorithm. IEEE Journal on Selected Areas in Communications, 16(2):140?152. [6] T. J. Richardson. (2000) The geometry of turbo-decoding dynamics. IEEE Transactions on Information Theory, 46(1):9?23. [7] T. Tanaka. (2001) Information geometry of mean-field approximation. In M. Opper and D. Saad, editors, Advanced Mean Field Methods ? Theory and Practice, chapter 17, pages 259?273. The MIT Press. [8] T. Tanaka, S. Ikeda, and S. Amari. (2002) Information-geometrical significance of sparsity in Gallager codes. in T. G. Dietterich et al. (eds.), Advances in Neural Information Processing Systems, vol. 14 (this volumn), The MIT Press. [9] J. S. Yedidia, W. T. Freeman, and Y. Weiss. (2001) Bethe free energy, Kikuchi approximations, and belief propagation algorithms. Technical Report TR2001?16, Mitsubishi Electric Research Laboratories.
2010 |@word version:2 nd:1 r:2 mitsubishi:1 bn:5 equimarginal:3 bc:1 bhattacharyya:1 wako:1 written:1 ikeda:3 selected:1 mpm:8 mathematical:4 direct:2 manner:1 brain:2 freeman:1 window:1 tlu:1 becomes:1 project:2 minimizes:1 thorough:1 every:3 collecting:1 exactly:1 control:1 positive:1 understood:2 local:1 analyzing:3 studied:1 unique:1 acknowledgment:1 practice:2 definite:1 sxt:1 procedure:1 area:1 projection:9 onto:4 marginalize:1 operator:1 put:1 equivalent:4 cjm:1 go:2 decomposable:1 correcting:2 stability:3 coordinate:1 elucidate:1 play:2 origin:1 approximated:3 updating:1 tributions:1 role:2 p5:3 calculate:2 observes:1 monograph:1 transforming:1 tr2001:1 dynamic:1 grateful:1 po:1 various:1 chapter:2 riken:2 univ:1 solve:1 amari:5 encoder:2 richardson:2 nagaoka:1 noisy:1 eigenvalue:1 kvw:2 product:1 loop:2 intuitive:3 convergence:1 optimum:2 converges:2 kikuchi:1 wider:1 derive:1 ac:2 eq:13 direction:1 tokyo:2 jst:1 shun:1 summation:1 correction:1 hold:9 sufficiently:2 equilibrium:11 lm:1 metropolitan:1 bsc:1 clearly:1 beth:2 always:1 mit:3 derived:1 focus:1 vk:1 improvement:2 check:1 tech:1 inst:1 relation:3 among:1 denoted:1 mackay:1 equal:1 field:3 identical:1 bnm:1 future:1 report:1 hni:1 fn1:5 divergence:1 replaced:2 geometry:6 connects:1 interest:1 investigate:1 adjust:1 hg:1 partial:2 taylor:1 theoretical:1 instance:1 metro:1 exchanging:3 maximization:1 loopy:4 cost:3 subset:1 saitama:1 uniform:1 submanifold:11 density:1 physic:1 decoding:38 connectivity:1 satisfied:1 american:1 japan:3 potential:1 summarized:3 coding:1 includes:6 satisfy:1 eei:1 depends:1 view:2 lot:3 accuracy:2 characteristic:4 who:1 toshiyuki:1 zy:2 kabashima:2 submitted:1 ed:1 definition:2 energy:4 proof:2 proved:1 higher:3 wei:1 strongly:2 furthermore:1 correlation:3 mceliece:2 propagation:7 defines:1 reveal:1 believe:1 dietterich:1 true:5 bsi:1 memoryless:1 symmetric:1 iteratively:1 laboratory:1 coincides:2 generalized:1 gg:1 geometrical:9 ef:1 common:2 ji:1 jp:3 volume:1 belong:1 approximates:1 marginals:1 shiro:2 aq:2 stable:1 attracting:1 curvature:3 posterior:3 codeword:2 binary:1 vt:4 transmitted:1 additional:1 berrou:1 fl1:4 eji:1 technical:1 calculation:1 cross:1 basic:2 expectation:6 yoshiyuki:1 proposal:1 background:2 schematically:1 saad:3 sr:58 probably:1 call:1 near:2 easy:1 enough:1 marginalization:11 gave:1 fukuoka:1 intensive:1 ultimate:1 hessian:1 cause:1 generally:3 useful:1 sign:1 estimated:4 vol:1 ichi:1 utilize:1 mystery:2 family:2 bit:3 convergent:2 cheng:1 turbo:33 bp:11 flat:8 bpo:1 generates:1 kyushu:1 taken:1 equation:4 chiranjib:1 mechanism:2 wvu:1 tractable:2 serf:1 presto:1 rewritten:2 yedidia:1 apply:1 v2:1 substitute:1 include:1 completed:1 opportunity:1 marginalized:1 neglect:1 k1:1 especially:1 society:1 flipping:1 diagonal:2 thank:1 decoder:10 manifold:1 assuming:1 code:13 index:1 difficult:2 unfortunately:1 negative:1 design:1 extended:1 communication:2 perturbation:2 kl:1 extensive:1 tap:1 pearl:2 tanaka:5 wy:1 sparsity:1 wz:1 belief:14 critical:1 advanced:2 lk:3 prior:2 understanding:6 marginalizing:1 embedded:1 viewpoint:2 editor:2 translation:1 surprisingly:1 parity:3 free:4 dis:1 hachioji:1 taking:1 face:1 opper:2 valid:1 kyutech:1 projected:1 transaction:3 approximate:1 global:1 reveals:1 receiver:1 don:1 iterative:6 sk:1 channel:3 bethe:1 expansion:1 necessarily:3 electric:1 factorizable:2 significance:1 main:2 n2:1 fig:4 exponential:1 theorem:10 specific:1 xt:18 normalizing:1 intractable:2 adding:1 jnk:2 sx:1 saddle:2 gallager:2 expressed:1 corresponds:1 satisfies:1 glavieux:1 goal:1 identity:1 room:1 fisher:3 feasible:1 included:2 principal:1 called:1 experimental:1 support:1 evaluate:1
1,109
2,011
Orientational and geometric determinants place and head- Neil Burgess & Tom Hartley Institute of Cognitive Neuroscience & Department of Anatomy, UCL 17 Queen Square, London WCIN 3AR, UK n. burgess@ucl.ac.uk. t.hartley@ucl.ac.uk Abstract We present a model of the firing of place and head-direction cells in rat hippocampus. The model can predict the response of individual cells and populations to parametric manipulations of both geometric (e.g. O'Keefe & Burgess, 1996) and orientational (Fenton et aI., 2000a) cues, extending a previous geometric model (Hartley et al., 2000). It provides a functional description of how these cells' spatial responses are derived from the rat's environment and makes easily testable quantitative predictions. Consideration of the phenomenon of remapping (Muller & Kubie, 1987; Bostock et aI., 1991) indicates that the model may also be consistent with nonparametric changes in firing, and provides constraints for its future development. 1 Introduction 'Place cells' recorded in the hippocampus of freely moving rats encode the rat's current location (O'Keefe & Dostrovsky, 1971; Wilson & McNaughton, 1993). In open environments a place cell will fire whenever the rat enters a specific portion of the environment (the 'place field'), independent of the rat's orientation (Muller et aI., 1994). This location-specific firing appears to be present on the rat's first visit to an environment (e.g. Hill, 1978), and does not depend on the presence of local cues such as odors on the floor or walls. The complementary pattern of firing has also been found in related brain areas: 'head-direction cells' that fire whenever the rat faces in a particular direction independent of its location (Taube et aI., 1990). Experiments involving consistent rotation of cues at or beyond the edge of the environment (referred to as 'distal' cues) produce rotation of the entire place (O'Keefe & Speakman, 1987; Muller et aI., 1987) or head-direction (Taube et aI., 1990) cell representation. Rotating cues within the environment does not produce this effect (Cressant et aI., 1997). Here we suggest a predicitive model of the mechanisms underlying these spatial responses. 2 Geometric influences given consistent orientation Given a stable directional reference (e.g. stable distal cues across trials), fields are determined by inputs tuned to detect extended obstacles or boundaries at particular bearings. That is, they respond whenever a boundary or obstacle occurs at a given distance along a given allocentric direction, independent of the rat's orientation. These inputs are referred to below as putative 'boundary vector cells' (BVCs). The functional form of these inputs has been estimated by recording from the same place cell in several environments of differing geometry within the same set of distal orientation cu~s (O'Keefe & Burgess, 1996; Hartley et al., 2000). That is, for a BVC i tuned to a boundary at distance di and bearing <Pi relative to the rat, the response to a houndary segment at distance r and bearing 9, subtending an angle cfJ at the rat, is given by: Cli == gi(r, fJ)CfJ, gi (r, fJ) ex: exp[-(r - d i )2/2a;ad(di )] V21r0";ad( d i) exp[-(fJ - <Pi)2 /2a~ng] X -----r===========--- (1) - /21ru V any 2 where the angular width aang is a constant but the radial width Urad == uo(1+di //3) so that the width of tuning to distance increases with the distance of peak response diD Constants 0"0 and /3 determine width at zero distance and its rate of increase with distance. The firing rate of BVC i, when the rat is at a location z, is found by integrating eli over (1 (this is done numerically as the distance r to the nearest boundary in direction fJ is a function of z, fJ and the geometry of the environment). A place cell's firing rate F(Z) is then simply the thresholded linear sum of the firing rates of the n Bves connected to it, Le. where e(z) is the Heaviside function (S(z) == z if x > 0; Sex) == 0 otherwise). All simulations have /3 == 183cm, Uo == 12.2cm, Urad == 0.2rad, while the threshold T can vary between simulations (e.g. between Figs. 1 and 2) but not between cells, and A is an arbitrary constant as absolute firing rates are not shown. Thus, in this model, a place cell's response is simply determined by the parameters d i and ifJi chosen for the set of BVes connected to it. Assuming a random selection of BVCs for each place cell, and a single value for T, the model provides a good fit to the characteristics of populations of place fields across different environments, such as the distribution of firing rates and field shapes and sizes. Inputs can also be chosen so as to fit a given place field so that its behavior in a new environment of different shape can be predicted. See Hartley et al. (2000) and Fig. 1. Like other models relying on the bearing to a landmark (Redish & Touretsky, 1996; McNaughton et al., 1996), the basic geometrical model assumes an accurate directionalreference, but does not state how this depends on the sensory input. Note that, as such, this model already captures effects of consistent rotation of orientation cues around an environment as a reorientation of the directional reference frame that in turn affects the directions along which BVCs are tuned to respond. Indeed, the effect of consistent rotation of orientation cues about a environment of fixed geometry is identical to the rotation of the environment within a fixed directional reference frame, and can be modelled in this way (see e.g. the square and diamond in Figs. 1b,c). 3 Model of geometric and orientation influences Models of head direction (Skaggs et al., 1995; Zhang, 1996) indicate how orientation might be derived. Internal inputs (e.g. vestibular or proprioceptive) maintain a consistent representation of heading within a ring of head-direction cells arranged to form a continuous attractor . Correlational learning of associations from visual inputs to head direction cells then allows the representation of head direction to be maintained in synch with the external world. These models account for the preferential influence of large cues at a stable bearing (i.e. at or beyond the edge of the environment), and effects of instability caused by continual movement of cues or disorientation of the rat. They also allow orientation to be maintained in the face of cue removal, unless all cues are removed in which case orientation is wholly reliant on internal inputs and will drift over time. In this paper we take a step towards providing a quantitative model for the combined influences of orientation cues and boundaries on the firing of place and head direction cells. Such a model should be able to predict the behaviour of these cells under arbitrary environmental manipulations, bearing in mind that some (extended) objects may be both orientation cues and boundaries. We focus on a series of experiments regarding inconsistent rotations of two extended cue cards (one white, one black) around the perimeter of a cylinder in the absence of any other orientation cues (Fenton et aI., 2000a). Each of these cards controls the orientation of the set of place fields when rotated together or alone (after removal of the other cue). When both are rotated inconsistently, place fields are displaced in a non-uniform manner, with the displacement of a field being a function of its location within the environment. These findings cannot be explained by a simple rotation of the reference frame. Fig. 2A shows how place fields are displaced following counter rotation of the two cue cards. Since the cue cards are orientation cues and also walls of the environment, explaining these data within the current framework requires two separate considerations: i) how the movement of the cards affects the BVC's directional reference frame, and ii) how the movement of the cards, acting as boundaries, directly affects the BVCs. We make the following assumptions: 1. The influence of a distal visual cue on the directional reference system is proportional to its proximity to the rat. 2. In the continued presence of color (or contrast) variation along a boundary to which a BVC responds, the BVC will become modulated by color: responding preferentially to, say, a white section of wall rather than the adjacent grey wall. In the absence of such variation it will revert to its unmodulated response. We note that assumption 1) is consistent with most implementations of the head direction model discussed above, in that the influence of an extended distal cue will increase with the angle subtended by it at the rat. We also note that assumption 2) implies the presence of synaptic learning (something not required by the rest of the model), albeit outside of the hippocampus. To avoid having to simulate enough random selections of BVCs to produce place fields at all locations within the environment and with all combinations of distance, bearing and color preferences, the model must be further simplified. To model the effect of cue manipulation on a place field in a location from which there are two cue cards at distances D i and bearings qli, we simulate a place cell for that location which receives inputs from two BVCs tuned to the distances D i and bearings qli, and to the most common color of boundary segments to which it respondes (across all positions of the rat). That is, di .= D i and 4>i == q>i in equation 1. For each location in the environment, we compute the shape of the place field formed by the thresholded sum of these BVCs, before and after the cue card manipulation. This simplification is broadly representative of the qualitative effect of the manipulation on the locations of place fields!. How does this model campare to the Fenton et aI. data? First we note that (due to assumption 1) each cue card can control the overall orientation of the place and head-direction representations. Similarly removing a cue card will have little effect, save for a slight rotation and/or transverse spreading of the Bve that responds to it (as it is no longer constrained by the color boundary, see assumption 2). When the cues are rotated inconsistently, the firing fields of the BVCs move relative to each other. The net effect of this on place fields and their centroids (Fig. 2B) compares well with the data (Fig. 2A) and is composed of two separate effects. First, the rotation of the cues produces a non-uniform distortion of the head direction system. The extent of rotation depends on the location of the animal relative to the cues as the closer a cue the more it affects the directional reference at that location (assumption 1) see Fig. 2C (ii). This distortion of the directional reference frame affects the orientations to which the boundary vector cells are tuned, and thus affects the location of place fields in an approximately rotational manner see Fig. 2C (iii).. Second, the movement of the cue cards directly affects the firing fields of the BVCs due to their color preferrence. This 'translational' effect is shown in Fig. 2C (iv). Note that neither translational nor rotational effects alone are sufficient to explain the observed data. Fenton et aI. (2000b) also make a distinction between translation and rotation in producing a phenomenological model of their data. However, as such, their model does not provide a mechanistic account at the level of cells, is specific to the cue-card manipulation they made and so does not make any prediction for head-direction cells or place cells in other experiments. 4 Non-parametric changes: 'remapping' Our model considers the pattern of firing of place cells when the rat is put into an environment of different shape, or when two very familiar landmarks are moved or removed. In these situations changes to patterns of firing tend to be parametric, and the model aims to capture the parametric relationships between firing pattern and environmental manipulation. However we note that, after several days or weeks of experience, the place cell representations of two environments of different shape gradually diverge (Lever et aI., 2002), such that the final representations can be said to have 'remapped' (MUller and Kubie, 1987). After 'remapping' a given cell might fire in only one of the environments, or might fire in both but in unrelated locations. Additionally, changing the color of the cue card in a grey cylinder from white to black can cause more rapid remapping such that the effect on the first day is probably best described as a slight rotation, with remapping occurring by the second day (Bostock et aI., 1991). Note that simply removing the cue card just causes the overall orientation of the place field representation to drift. Could the current model be extended to begin to understand these apparently nonparametric changes? The effects of replacing the cue card with a novel one are consistent with assumption 2 and the extra-hippocampal learning it implies: BVCs initially respond to the new color as they would upon removal of the cue card, with 1 Simulations of place fields with a larger number of BVCs indicate similar field movements, but of reduced magnitude in locations far from the cue cards. However the good match between the simple model and the data (Figs. 2A,B) suggests that the cue cards do provide the majority of BVC input. This might be due to learned salience over the extensive training period, and to the learning process implied by assumption 2. Against this, place fields formed by more the two BVC inputs (e.g. the four BVCs in Fig. Ic) generally give a better match to field shape, especially in locations far from the two cue cards. the slight rotation or spreading of the firing field noted above. Over time in the presence of the new color, the color modulation of BVCs sharpens such that those previously responding to white or grey no longer respond to black, while new BVCs that do respond to black begin to fire. Thus the original place fields (particularly those nearest the card and so most reliant on BVCs from that direction) will tend to fall below threshold, unless receiving a connection from a newly active blacksensitive BVC, in which case the field will change location. Equally, some previously silent place cells will become active due to input from a newly-active black-sensitive BVC. By contrast, the slow shape-dependent remapping would appear to require some additional mechanism. This may be related to the evidence of shorter-term learning of associations between place cells (M~hta et aI., 1997) or the NMDAdependent stability of place fields (Kentros et aI., 1998) or postulated processes of learned orthogonalisation of hippocampal representations (Marr, 1971; McClelland et aI., 1995; Treves & Rolls, 1992; Fuhs & Touretzky, 2000; Kali & Dayan, 2000). 5 Conclusion The model we have presented is consistent with a large body of detailed data on the effects of parametric environmental manipulations on place and head-direction cells. More importantly, it is a predictive model at the level of individual cells. Fig. 2C (ii) shows the prediction resulting from assumption 1) regarding the effect of the inconsistent cue card manipulation on head-direction cells. We note that there is an alternative to this location-dependent warping of head direction responses: a direction-dependent warping such that responses to north directions are tilted northwestwards while responses to south directions are tilted southwestwards. This would correspond to the alternative assumption that the influence of a distal visual cue on a head direction cell is proportional to the similarity of the average direction of the cue from the rat and the preferred direction of the cell. We chose to simulate the former (assumption 1) as this is consistent with current head-direction models in keeping the angular separation of preferred directions constant (but rotating all of them together as a function of the proximity of the rat to one or other cue card). The alternative assumption breaks this constancy, but would produce roughly equivalent results for place cell firing. Thus, on the basis of the Fenton et al. experiment on place cells we must predict one or other of the two effects on head-direction, or some combination of both. Beyond this, the model can predict the effect of essentially arbitrary parametric movements of cues and boundaries on place and head-direction 'cells over the short term. It also appears to be at least consistent with the nonparametric 'remapping' changes induced by color changes. Whether or not it can also predict the statistics of remapping over longer timescales in response to purely geometric changes is a question for future work. Acknowledgements: We thank John O'Keefe, Colin Lever and Bob Muller for many useful discussions. 6 References Bostock E, Muller RU, Kubie JL (1991) Experience-dependent modifications of hippocampal place cell firing Hippocampus 1, 193-206. Cressant A, Muller RU, Poucet B (1997) Failure of centrally placed objects to control the firing fields of hippocampal place cells. J. Neurosci. 17, 2531-2542. Fenton AA, Csizmadia G, & Muller RU (2000a). Conjoint control of hippocampal place cell firing by two visual stimuli. I. The effects of moving the stimuli on firing field positions. J. Gen. Physiol, 116, 191-209. Fenton AA, Csizmadia G, &. Muller RU (2000b). Conjoint control of hippocampal place cell firing by two visual stimuli. Ii. A vector-field theory that predicts modifications of the representation of the environment. J. Gen. Physiol, 116, 211-221. Fuhs MC, Touretzky DS (2000) Synaptic learning models of map separation in the hippocampus. Neurocomputing, 32:379-384. Hartley T, Burgess N, Lever C, Cacucci F, O'Keefe J (2000) Modeling place fields in terms of the cortical inputs to the hippocampus. Hippocampus, 10, 369-379. Hill AJ (1978) First occurrence of hippocampal spatial firing in a new environment. Exp. Neural 62, 282-297. Kali S, Dayan P (2000) The Involvement of Recurrent Connections in Area CA3 in Establishing the Properties of Place Fields: A Model. J. Neurosci. 20, 7463-7477. Kentros C, Hargreaves E, Hawkins RD, Kandel ER, Shapiro M, Muller RU (1998) Abolition of long-term stability of new hippocampal place cell maps by NMDA receptor blockade. Science, 280, 2121-2126. McNaughton BL, Knierim JJ, Wilson MA (1994) 'Vector encoding and the vestibular foundations of spatial cognition: a neurophysiological and computational hypothesis', In The Cognitive NeuroJJciences, (ed. Gazzaniga, M.) 585-596 (MIT Press, Boston, 1994). Lever CL, Wills T, Cacucci F, Burgess N, O'Keefe J (2002) Long-term plasticity in the hippocampal place cell representation of environmental geometry. Nature, in press. Marr D (1971) Simple memory: a theory for archicortex. Phil. Trans. Roy. Soc. Lond B 262, 23-81. McClelland JL, McNaughton BL, O'Reilly RC (1995) Why there are complementary learning-systems in the hippocampus and neocortex - insights from the successes and failures of connectionist models of learning and memory. Psychological Review 102, 419457. Mehta MR, Barnes CA, McNaughton BL (1997) Experience-dependent, asymmetric expansion of hippocampal place fields. Proc. Nat. Acad. Sci. 94, 8918-8921. Muller RU, Bostock E, Taube JS, Kubie JL (1994) On the directional firing properties of hippocampal place cells. J. Neurosci. 14 7235-7251. Muller RU, Kubie JL (1987) The effects of changes in the environment on the spatial firing of hippocampal complex-spike cells. J. Neurosci 7, 1951-1968. Muller RD, Kubie JL, Ranck JB (1987) Spatial firing patterns of hippocampal complexspike cells in a fixed environment. J. Neurosci., 7, 1935-1950. O'Keefe J, Burgess N (1996) Geometric Determinants of the Place Fields of Hippocampal Neurones. Nature 381, 425-428. O'Keefe J, Dostrovsky J (1971) The hippocampus as a spatial map: preliminary evidence from unit activity in the freely moving rat. Brain Res 34, 171-175. O'Keefe J, Speakman A (1987) Single unit activity in the rat hippocampus during a spatial memory task. Exp. Brain Res 68, 1-27. Redish AD, Touretzky DS (1996) Modeling interactions of the rat's place and head direction systems Advances in Neural Information Processing Systems, 8. D Touretzky, MC Mozer, ME Hasselmo (eds) pp. 61-67. MIT Press, Cambridge MA. Skaggs WE, Knierim JJ, Kudrimoti HS, McNaughton BL (1995) 'A model of the neural basis of the rat's sense of direction' Advances in Neural Information Processing Systems, 7. G Tesauro, D Touretzky &. TK Le'en (eds) pp. 51-58. MIT Press, Cambridge MA. Taube JS, Muller RD, Ranck JB (1990) Head-direction cells recorded from the postsubiculum in freely moving rats. 1. Description .and quantitative analysis. J. Neurosci 10, 420-435. Treves A, Rolls ET (1992) Computational constraints suggest the need for two distinct input systems to the hippocampal CA3 network. Hippocampus 2, 189-200. Wilson MA, McNaughton BL (1993) Dynamics of the hippocampal ensemble code for space. Science 261, 1055-1058. Zhang K (1996) Representation of spatial orientation by the intrinsic dynamics of the head-direction cell ensemble: a theory. J .Neurosci., 16, 2112-2126. b) e) Figure 1: Model of the geometrical influence on place fields (adapted from Hartley et aI., 2000), assuming a stable directional reference frame. Place fields are composed from thresholded linear sums of the firing rates of 'boundary vector cells' (BVCs). a) Above: Each BVC has a Gaussian tuned response to the presence of a boundary at a given distance and bearing from the rat (independent of its orientation). Below: The sharpness of tuning of a BVC decreases as the distance to which it is tuned increases. The only free parameters of a BVC are the distance and direction of peak response. b) Place fields recorded from the same cell in four environments of different shape or orientation relative to distal cues. c) Simulation of the place fields in b) by the best fitting set of 4 BVCs constrained to be in orthogonal directions (BVCs shown on the left, simulated fields on the right). The simulated cell can now be used to predict firing in novel situations. Real and predicted data from three novel environments are shown in d) and e) respectIvely, showing good qualitative agreement. A B =:==-~" k"" , i) iii) ii) c <It'" I \ - - :- =:==-"~. . ., --:q:::::'" ,=:.~ . ., ~ /' t f : i) '''t'-- ii) t ....... . ......~ iii) : ~ . :- ....-?3:;:/?? iv) Figure 2: Changes to place fields in a cylinder following inconsistent rotation of two cue cards. A) Experimental data shown in a birds-eye view of the cyclinder including the black and white cue-cards (adapted from Fenton et aI., 2000a). i) A place field with the cue cards in the 'standard' condition (used throughout training). ii) The place field after inconsistent rotation of each cue card by 12.5? further apart ('apart' condition)~ iii) The movement of the centroid of place field from the standard condition (tail of arrow) to the apart condition (head of arrow). B) Simulation of 21 place fields in the cyclinder in standard and apart conditions. Cue card locations are indicated by a black line (initial card positions are indicated by a dotted line to illustrate changes from one condition to another). i) and ii) show the place field nearest in location to that shown in A) in standard and apart conditions. iii) shows the movement of the centroids of simulated place fields between standard and apart conditions. C) i) Simulation of the movement of place field centroids between the standard and 'together' conditions (cue cards rotated 12.5? closer together). ii) The distortion of the preferred direction of a head direction cell. Arrows show the preferred direction in the 'apart' condition, the preferred direction was 'up' in the standard condition. iii) the movement of place field centroids between the standard and apart condition due solely to the directional distortion shown in ii). iv) the movement of place field centroids due solely to the movement of the cue cards acting as distinct cues (without any directional distortion shown in ii). The net effect of fields iii) and iv) is that shown in B iii).
2011 |@word h:1 trial:1 determinant:2 cu:1 sharpens:1 hippocampus:11 sex:1 open:1 mehta:1 grey:3 simulation:6 initial:1 bvc:12 series:1 tuned:7 ranck:2 current:4 must:2 john:1 tilted:2 physiol:2 plasticity:1 shape:8 alone:2 cue:54 short:1 cacucci:2 provides:3 location:20 preference:1 zhang:2 rc:1 along:3 become:2 qualitative:2 fitting:1 manner:2 indeed:1 rapid:1 roughly:1 behavior:1 nor:1 brain:3 relying:1 little:1 begin:2 underlying:1 unrelated:1 remapping:8 cm:2 differing:1 finding:1 quantitative:3 continual:1 uk:3 control:5 unit:2 uo:2 unmodulated:1 producing:1 appear:1 before:1 local:1 acad:1 receptor:1 encoding:1 establishing:1 firing:28 modulation:1 approximately:1 solely:2 might:4 black:7 chose:1 bird:1 suggests:1 speakman:2 cfj:2 kubie:6 displacement:1 wholly:1 area:2 reilly:1 radial:1 integrating:1 suggest:2 cannot:1 selection:2 put:1 influence:8 instability:1 equivalent:1 map:3 phil:1 sharpness:1 insight:1 continued:1 importantly:1 marr:2 population:2 stability:2 variation:2 mcnaughton:7 hypothesis:1 agreement:1 roy:1 particularly:1 asymmetric:1 predicts:1 observed:1 constancy:1 fuhs:2 enters:1 capture:2 connected:2 movement:12 removed:2 counter:1 decrease:1 mozer:1 environment:27 dynamic:2 depend:1 segment:2 predictive:1 purely:1 upon:1 basis:2 easily:1 revert:1 distinct:2 london:1 outside:1 larger:1 say:1 distortion:5 otherwise:1 statistic:1 neil:1 gi:2 final:1 orthogonalisation:1 net:2 ucl:3 interaction:1 gen:2 description:2 moved:1 cli:1 qli:2 wcin:1 extending:1 produce:5 ring:1 rotated:4 object:2 bve:1 tk:1 recurrent:1 ac:2 illustrate:1 nearest:3 soc:1 predicted:2 indicate:2 implies:2 direction:40 anatomy:1 hartley:7 require:1 behaviour:1 wall:4 preliminary:1 proximity:2 around:2 hawkins:1 ic:1 exp:4 cognition:1 predict:6 week:1 vary:1 proc:1 spreading:2 sensitive:1 hasselmo:1 kudrimoti:1 mit:3 gaussian:1 aim:1 rather:1 avoid:1 wilson:3 encode:1 derived:2 focus:1 indicates:1 contrast:2 centroid:6 detect:1 sense:1 dependent:5 dayan:2 entire:1 initially:1 overall:2 translational:2 orientation:21 kali:2 development:1 animal:1 spatial:9 constrained:2 field:49 having:1 ng:1 identical:1 future:2 connectionist:1 stimulus:3 jb:2 composed:2 neurocomputing:1 individual:2 familiar:1 geometry:4 fire:5 attractor:1 maintain:1 cylinder:3 perimeter:1 accurate:1 edge:2 closer:2 preferential:1 experience:3 shorter:1 orthogonal:1 unless:2 iv:4 rotating:2 re:2 psychological:1 dostrovsky:2 modeling:2 obstacle:2 ar:1 queen:1 subtending:1 ca3:2 uniform:2 combined:1 peak:2 receiving:1 diverge:1 together:4 lever:4 recorded:3 cognitive:2 external:1 fenton:8 account:2 redish:2 north:1 postulated:1 caused:1 ad:3 depends:2 break:1 view:1 apparently:1 portion:1 poucet:1 square:2 formed:2 roll:2 characteristic:1 ensemble:2 correspond:1 directional:11 modelled:1 mc:2 bob:1 explain:1 touretzky:5 whenever:3 synaptic:2 ed:3 against:1 failure:2 pp:2 di:4 newly:2 color:11 orientational:2 nmda:1 appears:2 day:3 tom:1 response:13 arranged:1 done:1 angular:2 just:1 subtended:1 d:2 receives:1 replacing:1 aj:1 indicated:2 effect:20 former:1 proprioceptive:1 white:5 distal:7 adjacent:1 blockade:1 during:1 width:4 maintained:2 noted:1 rat:25 hippocampal:16 hill:2 fj:5 geometrical:2 consideration:2 novel:3 archicortex:1 common:1 rotation:16 functional:2 jl:5 association:2 discussed:1 slight:3 tail:1 numerically:1 cambridge:2 ai:17 tuning:2 rd:3 postsubiculum:1 similarly:1 phenomenological:1 moving:4 stable:4 longer:3 similarity:1 something:1 j:2 involvement:1 apart:8 tesauro:1 manipulation:9 success:1 muller:14 additional:1 floor:1 mr:1 freely:3 determine:1 taube:4 period:1 colin:1 ii:11 match:2 long:2 equally:1 visit:1 prediction:3 involving:1 basic:1 essentially:1 cell:50 extra:1 rest:1 probably:1 south:1 recording:1 tend:2 induced:1 inconsistent:4 presence:5 iii:8 enough:1 affect:7 skaggs:2 burgess:7 fit:2 silent:1 regarding:2 whether:1 neurones:1 cause:2 jj:2 generally:1 useful:1 detailed:1 nonparametric:3 neocortex:1 mcclelland:2 reduced:1 shapiro:1 dotted:1 neuroscience:1 estimated:1 broadly:1 four:2 threshold:2 changing:1 neither:1 thresholded:3 sum:3 hta:1 angle:2 eli:1 respond:5 place:64 throughout:1 separation:2 putative:1 simplification:1 centrally:1 barnes:1 activity:2 adapted:2 constraint:2 simulate:3 lond:1 department:1 combination:2 across:3 modification:2 explained:1 gradually:1 equation:1 previously:2 turn:1 mechanism:2 mind:1 mechanistic:1 occurrence:1 save:1 alternative:3 odor:1 allocentric:1 original:1 inconsistently:2 assumes:1 responding:2 testable:1 especially:1 bl:5 implied:1 move:1 warping:2 already:1 question:1 occurs:1 spike:1 parametric:6 responds:2 said:1 distance:14 separate:2 card:30 thank:1 sci:1 landmark:2 majority:1 simulated:3 me:1 extent:1 considers:1 assuming:2 ru:8 touretsky:1 code:1 relationship:1 providing:1 rotational:2 preferentially:1 implementation:1 diamond:1 displaced:2 situation:2 extended:5 head:25 frame:6 arbitrary:3 bostock:4 knierim:2 drift:2 treves:2 transverse:1 required:1 extensive:1 connection:2 rad:1 distinction:1 learned:2 vestibular:2 trans:1 beyond:3 able:1 gazzaniga:1 below:3 pattern:5 remapped:1 including:1 memory:3 eye:1 review:1 geometric:7 acknowledgement:1 removal:3 relative:4 proportional:2 conjoint:2 foundation:1 sufficient:1 consistent:11 pi:2 translation:1 placed:1 keeping:1 free:1 heading:1 salience:1 allow:1 understand:1 institute:1 explaining:1 fall:1 face:2 absolute:1 boundary:15 cortical:1 world:1 sensory:1 made:1 simplified:1 reorientation:1 far:2 preferred:5 active:3 continuous:1 why:1 additionally:1 nature:2 ca:1 expansion:1 bearing:10 cl:1 complex:1 synch:1 did:1 timescales:1 neurosci:7 arrow:3 complementary:2 body:1 fig:12 referred:2 representative:1 en:1 slow:1 position:3 kandel:1 removing:2 specific:3 showing:1 er:1 evidence:2 intrinsic:1 albeit:1 keefe:10 magnitude:1 nat:1 occurring:1 boston:1 simply:3 reliant:2 visual:5 neurophysiological:1 aa:2 environmental:4 ma:4 towards:1 absence:2 change:11 determined:2 acting:2 correlational:1 experimental:1 internal:2 modulated:1 heaviside:1 phenomenon:1 ex:1
1,110
2,012
Optimising Synchronisation Times for Mobile Devices Neil D. Lawrence Department of Computer Science, Regent Court, 211 Portobello Road, Sheffield, Sl 4DP, U.K. neil~dcs.shef . ac.uk Antony 1. T. Rowstron Christopher M . Bishop Michael J. Taylor Microsoft Research 7 J. J. Thomson Avenue Cambridge, CB3 OFB, U.K. {antr,cmbishop,mitaylor}~microsoft.com Abstract With the increasing number of users of mobile computing devices (e.g. personal digital assistants) and the advent of third generation mobile phones, wireless communications are becoming increasingly important. Many applications rely on the device maintaining a replica of a data-structure which is stored on a server, for example news databases, calendars and e-mail. ill this paper we explore the question of the optimal strategy for synchronising such replicas. We utilise probabilistic models to represent how the data-structures evolve and to model user behaviour. We then formulate objective functions which can be minimised with respect to the synchronisation timings. We demonstrate, using two real world data-sets, that a user can obtain more up-to-date information using our approach. 1 Introduction As the available bandwidth for wireless devices increases, new challenges are presented in the utilisation of such bandwidth. Given that always up connections are generally considered infeasible an important area of research in mobile devices is the development of intelligent strategies for communicating between mobile devices and servers. ill this paper we consider the scenario where we are interested in maintaining, on a personal digital assistant (PDA) with wireless access, an up-to-date replica of some, perhaps disparate, data-structures which are evolving in time. The objective is to make sure our replica is not 'stale'. We will consider a limited number of connections or synchronisations. Each synchronisation involves a reconciliation between the replica on the mobile device and the data-structures of interest on the server. Later in the paper we shall examine two typical examples of such an application,an internet news database and a user's e-mail messages. Currently the typical strategy! for performing such reconciliations is to synchronise every M minutes, lSee, for example, AvantGo http://vvv. avantgo. com. where M is a constant, we will call this strategy the uniformly-spaced strategy. We will make the timings of the synchronisations adaptable by developing a cost function that can be optimised with respect to the timings, thereby improving system performance. 2 Cost Function We wish to minimise the staleness of the replica, where we define staleness as the time between an update of a portion of the data-structure on the server and the time of the synchronisation of that update with the PDA. For simplicity we shall assume that each time the PDA synchronises all the outstanding updates are transferred. Thus, after synchronisation the replica on the mobile device is consistent with the master copy on the server. Therefore, if skis the time of the kth synchronisation in a day, and updates to the data-structure occur at times Uj then the average staleness of the updates transferred during synchronisation Sk will be (1) As well as staleness, we may be interested in optimising other criteria. For example, mobile phone companies may seek to equalise demand across the network by introducing time varying costs for the synchronisations, c(t). Additionally one could argue that there is little point in keeping the replica fresh during periods when the user is unlikely to check his PDA, for example when he or she is sleeping. We might therefore want to minimise the time between the user's examination of the PDA and the last synchronisation. If the user looks at the PDA at times ai then we can express this as (2) Given the timings Uj and ai, the call cost schedule c(t) and K synchronisations, the total cost function may now be written K C= L (-aFk + fJSk + C(Sk)) ' (3) k=l where a and fJ are constants with units of ~~~:y which express the relative importance of the separate parts of the cost function. Unfortunately, of course, whilst we are likely to have knowledge of the call cost schedule, c(t), we won't know the true timings {Uj} and {ai} and the cost function will be a priori incomputable. If, though, we have historic data2 relating to these times, we can seek to make progress by modelling these timings probabilistically. Then, rather than minimising the actual cost function, we can look to minimise the expectation of the cost function under these probabilistic models. 3 Expected Cost There are several different possibilities for our modelling strategy. A sensible assumption is that there is independence between different parts of the data-structure (i.e. e-mail and business news can be modelled separately), however, there may be dependencies between update times which occur within the same part. The 2When modelling user ru:cess times, if historic data is not available, models could also be constructed by querying the user about their likely ru:tivities. periodicity of the data may be something we can take advantage of, but any nonstationarity in the data may cause problems. There are various model classes we could consider; for this work however, we restrict ourselves to stationary models, and ones in which updates arrive independently and in a periodic fashion . . Let us take T to be the largest period of oscillation in the data arrivals, for a particular portion of a data-structure. We model this portion with a probability distribution, Pu(t). Naturally more than one update may occur in that interval, therefore our probability distribution really specifies a distribution over time given one that one update (or user access) has occurred. To fully specify the model we also are required to store the expected number of updates, Ju , (or accesses, J a ) that occur in that interval. The expected value of Sk may now be written, (4) where Op(:v) is an expectation under the distribution p(x), Au(t) = JuPu(t) can be viewed as the rate at which updates are occurring and So = SK - T. We can model the user access times, ai, in a similar manner, which leads us to the expected value of the freshness, (Fk)Pa(t) = kk +l Aa(t)(t - sk)dt, where Aa(t) = JaPa(t) The overall expected cost, which we will utilise as our objective function, may therefore be written J: K (C) = L (Sk)p" - (Fk)Pa + C(Sk)) . (5) k=l 3.1 Probabilistic Models. We now have an objective function which is a function of the variables we wish to optimise, the synchronisation times, but whilst we have mentioned some characteristics of the models Pu(t) and Pa(t) we have not yet fully specified their form. We have decreed that the models should be periodic and that they may consider each datum to occur independently. In effect we are modelling data which is mapped to a circle. Various options are available for handling such models; for this work, we constrain our investigations to kernel density estimates (KDE). In order to maintain periodicity, we must select a basis function for our KDE which represents a distribution on a circle, one simple way of achieving this aim is to wrap a distribution that is defined along a line to the circle (Mardia, 1972). A traditional density which represents a distribution on the line, p(t), may be wrapped around a circle of circumference T to give us a distribution defined on the circle, p( 0), where 0 = t mod T. This means a basis function with its centre at T - 8, that will typically have probability mass when u > T, wraps around to maintain a continuous density at T. The wrapped Gaussian distribution 3 that we make use of takes the form (6) The final kernel density estimate thus consists of mapping the data points tn -t On 3In practice we must approximate the wrapped distribution by restricting the number of terms in the sum. "-' ~60 .? ~40 "-' .S 1)520 ~ OJ H Co> OJ ""0 Thu Fri Sat Sun Thu Fri Sat Sun ~ -20 Figure 1: Left: part of the KDE developed for the business category together with a piecewise constant approximation. Middle: the same portion of the KDE for the FA Carling Premiership data. Right: percent decrease in staleness vs number of synchronisations per day for e-mail data. and obtaining a distribution 1 N p(()) = N L W N(()I()n, (}"2), (7) n=l where N is the number of data-points and the width parameters, (}", can be set through cross validation. Models of this type may be made use of for both Pu(t) and Pa(t) . 3.2 Incorporating Prior Knowledge. The underlying component frequencies of the data will clearly be more complex than simply a weekly or daily basis. Ideally we should be looking to incorporate as much of our prior knowledge about these component frequencies as possible. IT we were modelling financial market's news, for example, we would expect weekdays to have similar characteristics to each other, but differing characteristics from the weekend. For this work, we considered four different scenarios of this type. For the first scenario, we took T = 1 day and placed no other constraints on the model. For the second we considered the longest period to be one week, T = 1 week, and placed no further constraints on the model. For the remaining two though we also considered T to be one week, but we implemented further assumptions about the nature of the data. Firstly we split the data into weekdays and weekends. We then modelled these two categories separately, making sure that we maintained a continuous function for the whole week by wrapping basis functions between weekdays and weekends. Secondly we split the datainto weekdays, Saturdays and Sundays, modelling each category separately and again wrapping basis functions across the days. 3.3 Model Selection. To select the basis function widths, and to determine which periodicity assumption best matched the data, we ?utilised ten fold cross validation. For each different periodicity we used cross validation to first select the basis function width. We then compared the average likelihood across the ten validation sets, selecting the periodicity with the highest associated value. 4 Optimising the Synchronisation Times Given that our user model, Pa(t), and our data model, Pu(t) will be a KDE based on wrapped Gaussians, we should be in a position to compute the required integrals '" C/J Q) Q) t{360 >=1 t{360 >=1 <il .....,40 X X C/J + )I( .S ]40 C/J .S S520 S520 ~ ~ Q) Q) !-< !-< U ~ -20 X ~ C/J C/J Q) !I ~ ....., 4 X "0 ~ j ~ C/J Q) 2 2 X X -20 Figure 2: Results from the news database tests. Left: February /March based models tested on April. Middle: March/ April testing on May. Right: April/May testing on June. The results are in the form of box plots. The lower line of the box represents the 25th percentile of the data, the upper line the 75th percentile and the central line the median. The 'whiskers' represent the maximum extent of the data up to 1.5 x (75th percentile - 25th percentile). Data which lies outside the whiskers is marked with crosses. .S X 12 Q) C/J ~ ~20 u + + 40 >=1 .$20 U Q) "0 60 x !i I~! 24 X X xx Q) "0 X X X ~40 X X -60 + xx X xXx in (5) and evaluate our objective function and derivatives thereof. First though, we must give some attention to the target application for the algorithm. A known disadvantage of the standard kernel density estimate is the high storage requirements of the end model. The model requires that N floating point numbers must be stored, where N is the quantity of training data. Secondly, integrating across the cost function results in an objective function which is dependent on a large number of evaluations of the cumulative Gaussian distribution. Given that we envisage that such optimisations could be occurring within a PDA or mobile phone, it would seem prudent to seek a simpler approach to the required minimisation. An alternative approach that we explored is to approximate the given distributions with a functional form which is more amenable to the integration. For example, a piecewise constant approximation to the KDE simplifies the integral considerably. It leads to a piecewise constant approximation for Aa(t) and Au(t). Integration over which simply leads to a piecewise linear function which may be computed in a straightforward manner. Gradients may also be computed. We chose to reduce the optimisation to a series of one-dimensional line minimisations. This can be achieved in the following manner. First, note that the objective function, as a function of a particular synchronisation time Sk, may be written: (8) In other words, each synchronisation is only dependent on that of its neighbours. We may therefore perform the optimisation by visiting each synchronisation time, Sk, in a random order and optimising its position between its neighbours, which involves a one dimensional line minimisation of (8). This process, which is guaranteed to find a (local) minimum in our objective function, may be repeated until convergence. 5 Results In this section we mainly explore the effectiveness of modelling the data-structures of interest. We will briefly touch upon the utility of modelling the cost evolution and user accesses in Section 5.2 but we leave a more detailed exploration of this area to later works. 5.1 Modelling Data Structures To determine the effectiveness of our approach, we utilised two different sources of data: a news web-site and e-mail on a mail server. The news database data-set was collected from the BBC News web site4 . This site maintains a database of articles which are categorised according to subject, for example, UK News, Business News, Motorsport etc .. We had six months of data from February to July 2000 for 24 categories of the database. We modelled the data by decomposing it into the different categories and modelling each separately. This allowed us to explore the periodicity of each category independently. This is a sensible approach given that the nature of the data varies considerably across the categories. .Two extreme examples are Business news and FA Carling Premiership news 5 , Figure 1. Business news predominantly arrives during the week whereas FA Carling Premiership news arrives typically just after soccer games finish on a Saturday. Business news was best modelled on a Weekday/Weekend basis, and FA Carling Premiership news was best modelled on a Weekday /Saturday /Sunday basis. To evaluate the feasibility of our approach, we selected three consecutive months of data. The inference step consisted of constructing our models on data from the first two months. To restrict our investigations to the nature of the data evolution only, user access frequency was taken to be uniform and cost of connection was considered to be constant. For the decision step we considered 1 to 24 synchronisations a day. The synchronisation times were optimised for each category separately, they were initialised with a uniformly-spaced strategy, optimisation of the timings then proceeded as described in Section 4. The staleness associated with these timings was then computed for the third month. This value was compared with the staleness resulting from the uniformly-spaced strategy containing the same number of synchronisations 6 . The percentage decrease in staleness is shown in figures 2 and 3 in the form of box-plots. 60 x x x 40 x x 20 12 -20 X X 00 g}40 ? '"@ t260 .S Xx x ~XX ~ x ~,(?<: ~x ~ x \xx ++ .,? + )( + x x x x x + + + + + -120 + + -140 + + + -160 + + + + -180 + -200 Figure 3: May/June based models tested on July. + signifies the FA Carling Premiership Stream. Generally an improvement in performance is observed, however, we note that in Figure 3 the performance for several categories is extremely 4http://news.bbc.co.uk. 5The FA Carling Premiership is England's premier division soccer. 6The uniformly-spaced strategy's staleness varies with the timing of the first of the K synchronisations. This figure was therefore an average of the staleness from all possible starting points taken at five minute intervals. poor. In particular the FA Carling Premiership stream in Figure 3. The poor performance is caused by the soccer season ending in May. As a result relatively few articles are written in July, most of them concerning player transfer speculation, and the timing of those articles is very different from those in May. In other words the data evolves in a non-stationary manner which we have not modelled. The other poor performers are also sports related categories exhibiting non-stationarities. The e-mail data-set was collected by examining the logs of e-mail arrival times for 9 researchers from Microsoft's Cambridge research lab. This data was collected for January and February 2001. We utilised the January data to build the probabilistic models and the February data to evaluate the average reduction in staleness. Figure 1 shows the results obtained. In practice, a user is more likely to be interested in a combination of different categories of data. Perhaps several different streams of news and his e-mail. Therefore, to recreate a more realistic situation where a user has a combination of interests, we also collected e-mail arrivals for three users from February, March and April 2000. We randomly generated user profiles by sampling, without replacement, five categories from the available twenty-seven, rejecting samples where more than one e-mail stream was selected. We then modelled the users' interests by constructing an unweighted mixture of the five categories and proceeded to optimise the synchronisation times based on this model. This was performed one hundred times. The average staleness for the different numbers of synchronisations per day is shown in Figure 4. Note that the performance for the combined categories is worse than it is for each individually. This is to be expected as the entropy of the combined model will always be greater than that of its constituents, we therefore have less information about arrival times, and as a result there are less gains to be made over the uniformlyspaced strategy7. 5.2 Affect of Cost and User Model In the previous sections we focussed on modelling the evolution of the databases. Here we now briefly turn our attention to the other portions of the system, user behaviour and connection cost. For this preliminary study, it proved difficult to obtain high quality data representing user access times. We therefore artificially generated a model which represents a user who accesses there device frequently at breakfast, lunchtime and during the evening, and rarely at night. Figure 4 simply shows the user model, Pa(t), along with the result of optimising the cost function for uniform data arrivals and fixed cost under this user model. Note how synchronisation times are suggested just before high periods of user activity are about to occur. Also in Figure 4 is the effect of a varying cost, c(t), under uniform Pa(t) and Pa(t). Currently most mobile internet access providers appear to be charging a flat fee for call costs (typically in the U.K. about 15 cents per minute). However, when demand on their systems rise they may wish to incorporate a varying cost to flatten peak demands. This cost could be an actual cost for the user, or alternatively a 'shadow price' specified by service provider for controlling demand (Kelly, 2000). We give a simple example of such a call cost in Figure 4. For this we considered user access and data update rates to be constant. Note how the times move away from periods of high cost. 7The uniformly-spaced strategy can be shown to be optimal when the entropy of the underlying distribution is maximised (a uniform distribution across the interval). "-' gj 60 >=1 Cl) 7Lo 0.3 0.25 ....., 0.2 "-' 1200 ~ 900 ~20 U 0.15 '...." Cl) :::::::: 600 '" 0.1 U 0.05 0 00:00 U Cl) 300 "'0 ~ 08:00 X .S 00:00 08:00 16:00 00:00 X -20 Figure 4: Left: change in synchronisation times for variable user access rates. x shows the initialisation points, + the end points. Middle: change in synchronisation times for a variable cost. Right: performance improvements for the combination of news and e-mail. 6 Discussion The optimisation strategy we suggest could be sensitive to local minima, we did not try a range of different initialisations to explore this phenomena. However, by initialising with the uniformly-spaced strategy we ensured that we increased the objective function relative to the standard strategy. The month of July showed how a non-stationarity in the data structure can dramatically affect our performance. We are currently exploring on-line Bayesian models which we hope will track such non-stationarities. The system we have explored in this work assumed that the data replicated on the mobile device was only modified on the server. A more general problem is that of mutable replicas where the data may be modified on the server or the client. Typical applications of such technology include mobile databases, where sales personnel modify portions of the database whilst on the road, and a calendar application on a PDA, where the user adds appointments on the PDA. Finally there are many other applications of this type of technology beyond mobile devices. Web crawlers need to estimate when pages are modified to maintain a representative cache (eho and Garcia-Molina, 2000). Proxy servers could also be made to intelligent maintain their caches of web-pages up-to-date (Willis and Mikhailov, 1999; Wolman et al., 1999) . References Cho, J. and H. Garcia-Molina (2000). Synchronizing a database to improve freshness. In Proceedings 2000 ACM International Conference on Management of Data (SIG- MOD). Kelly, F. P. (2000). Models for a self-managed internet. Philosophical Transactions of the Royal Society A358, 2335-2348. Mardia, K. V. (1972). Statistics of Directional Data. London: Academic Press. Rowstron, A. 1. T., N. D. Lawrence, and C. M. Bishop (2001). Probabilistic modelling of replica divergence. In Proceedings of the 8th Workshop on Hot Topics in Operating Systems HOTOS (VIII). Willis, C. E. and M. Mikhailov (1999). Towards a better understanding of web resources and server responses for improved caching. In Proceedings of the 8th International World Wide Web Conference, pp. 153-165. Wolman, A., G. M. Voelker, N. Sharma, N. Cardwell, A. Karlin, and H. M. Levy (1999). On the scale and performance of co-operative web proxy caching. In 17th ACM Symposium Operating System Principles (SOSP'99), pp. 16-3l. Yu, H. and A. Vahdat (2000). Design and evaluation of a continuous consistency model for replicated services. In 4th Symposium on Operating System Design and Implementation (OSDI).
2012 |@word proceeded:2 briefly:2 middle:3 seek:3 weekday:6 thereby:1 reduction:1 series:1 selecting:1 initialisation:2 com:2 yet:1 written:5 must:4 realistic:1 plot:2 update:12 v:1 stationary:2 selected:2 device:11 maximised:1 data2:1 firstly:1 simpler:1 five:3 along:2 constructed:1 saturday:3 symposium:2 consists:1 regent:1 manner:4 market:1 expected:6 examine:1 frequently:1 company:1 little:1 actual:2 cache:2 increasing:1 xx:5 underlying:2 matched:1 mass:1 advent:1 developed:1 whilst:3 differing:1 cmbishop:1 synchronisation:27 every:1 stationarities:2 weekly:1 ensured:1 uk:3 sale:1 unit:1 appear:1 before:1 service:2 timing:10 local:2 modify:1 vahdat:1 optimised:2 becoming:1 might:1 chose:1 au:2 co:3 limited:1 range:1 testing:2 practice:2 area:2 evolving:1 word:2 road:2 integrating:1 flatten:1 suggest:1 selection:1 storage:1 circumference:1 straightforward:1 attention:2 starting:1 independently:3 formulate:1 simplicity:1 communicating:1 his:2 financial:1 lsee:1 target:1 controlling:1 user:30 sig:1 pa:8 database:10 observed:1 news:19 sun:2 decrease:2 highest:1 mentioned:1 pda:9 ideally:1 rowstron:2 personal:2 upon:1 bbc:2 division:1 basis:9 lunchtime:1 various:2 weekend:4 london:1 outside:1 voelker:1 calendar:2 statistic:1 neil:2 envisage:1 final:1 advantage:1 took:1 karlin:1 date:3 constituent:1 convergence:1 requirement:1 leave:1 ac:1 op:1 progress:1 implemented:1 involves:2 shadow:1 exhibiting:1 exploration:1 behaviour:2 really:1 investigation:2 preliminary:1 secondly:2 exploring:1 around:2 considered:7 lawrence:2 mapping:1 week:5 consecutive:1 assistant:2 currently:3 sensitive:1 individually:1 largest:1 hope:1 clearly:1 always:2 gaussian:2 aim:1 sunday:2 rather:1 modified:3 caching:2 season:1 mobile:13 varying:3 probabilistically:1 minimisation:3 june:2 she:1 longest:1 modelling:12 check:1 likelihood:1 mainly:1 improvement:2 afk:1 inference:1 osdi:1 dependent:2 unlikely:1 typically:3 interested:3 overall:1 ill:2 prudent:1 priori:1 development:1 integration:2 sampling:1 optimising:5 represents:4 synchronizing:1 look:2 yu:1 intelligent:2 piecewise:4 few:1 randomly:1 neighbour:2 divergence:1 floating:1 ourselves:1 replacement:1 microsoft:3 maintain:4 stationarity:1 interest:4 message:1 possibility:1 evaluation:2 mixture:1 extreme:1 arrives:2 amenable:1 integral:2 daily:1 taylor:1 circle:5 increased:1 disadvantage:1 signifies:1 cost:27 introducing:1 uniform:4 hundred:1 examining:1 stored:2 dependency:1 varies:2 antony:1 periodic:2 considerably:2 combined:2 cho:1 ju:1 density:5 peak:1 international:2 probabilistic:5 minimised:1 michael:1 together:1 again:1 central:1 management:1 containing:1 worse:1 derivative:1 caused:1 stream:4 later:2 performed:1 utilised:3 lab:1 try:1 thu:2 portion:6 option:1 maintains:1 il:1 characteristic:3 who:1 spaced:6 directional:1 modelled:7 bayesian:1 rejecting:1 provider:2 researcher:1 premier:1 nonstationarity:1 frequency:3 initialised:1 pp:2 thereof:1 naturally:1 associated:2 gain:1 proved:1 knowledge:3 schedule:2 adaptable:1 dt:1 day:6 xxx:1 breakfast:1 specify:1 response:1 april:4 improved:1 though:3 box:3 just:2 until:1 web:7 christopher:1 touch:1 night:1 quality:1 perhaps:2 stale:1 effect:2 consisted:1 true:1 managed:1 evolution:3 staleness:12 utilisation:1 wrapped:4 during:4 width:3 game:1 self:1 maintained:1 portobello:1 percentile:4 soccer:3 won:1 criterion:1 thomson:1 demonstrate:1 willis:2 tn:1 fj:1 percent:1 predominantly:1 functional:1 he:1 occurred:1 relating:1 cambridge:2 ai:4 fk:2 consistency:1 centre:1 had:1 access:11 operating:3 gj:1 etc:1 pu:4 add:1 something:1 showed:1 phone:3 scenario:3 store:1 server:10 freshness:2 molina:2 minimum:2 greater:1 performer:1 fri:2 determine:2 period:5 sharma:1 july:4 reconciliation:2 england:1 academic:1 minimising:1 cross:4 concerning:1 feasibility:1 sheffield:1 optimisation:5 expectation:2 represent:2 kernel:3 achieved:1 sleeping:1 whereas:1 want:1 separately:5 shef:1 interval:4 appointment:1 operative:1 median:1 source:1 sure:2 subject:1 vvv:1 mod:2 seem:1 effectiveness:2 call:5 split:2 independence:1 finish:1 affect:2 bandwidth:2 restrict:2 reduce:1 simplifies:1 avenue:1 court:1 minimise:3 recreate:1 six:1 utility:1 cause:1 dramatically:1 generally:2 detailed:1 ten:2 category:14 http:2 specifies:1 sl:1 percentage:1 per:3 track:1 shall:2 express:2 four:1 achieving:1 cb3:1 ce:1 replica:10 sum:1 master:1 arrive:1 oscillation:1 decision:1 initialising:1 fee:1 internet:3 guaranteed:1 datum:1 fold:1 activity:1 occur:6 constraint:2 constrain:1 flat:1 extremely:1 performing:1 relatively:1 transferred:2 department:1 developing:1 according:1 march:3 poor:3 combination:3 across:6 increasingly:1 evolves:1 making:1 taken:2 resource:1 turn:1 know:1 end:2 available:4 gaussians:1 decomposing:1 away:1 alternative:1 cent:1 remaining:1 include:1 maintaining:2 eho:1 uj:3 build:1 february:5 society:1 objective:9 move:1 question:1 quantity:1 wrapping:2 strategy:13 fa:7 traditional:1 visiting:1 gradient:1 dp:1 kth:1 wrap:2 separate:1 mapped:1 sensible:2 seven:1 mail:12 argue:1 extent:1 collected:4 topic:1 fresh:1 viii:1 ru:2 kk:1 difficult:1 unfortunately:1 kde:6 disparate:1 rise:1 design:2 implementation:1 ski:1 twenty:1 perform:1 upper:1 january:2 situation:1 communication:1 looking:1 dc:1 ofb:1 required:3 specified:2 speculation:1 connection:4 philosophical:1 beyond:1 suggested:1 challenge:1 optimise:2 oj:2 royal:1 charging:1 hot:1 business:6 rely:1 examination:1 client:1 representing:1 improve:1 technology:2 prior:2 kelly:2 understanding:1 evolve:1 relative:2 fully:2 historic:2 expect:1 whisker:2 generation:1 querying:1 digital:2 validation:4 consistent:1 proxy:2 article:3 principle:1 lo:1 course:1 periodicity:6 placed:2 wireless:3 copy:1 keeping:1 infeasible:1 last:1 wide:1 focussed:1 world:2 cumulative:1 ending:1 unweighted:1 made:3 replicated:2 wolman:2 transaction:1 approximate:2 categorised:1 sat:2 assumed:1 alternatively:1 continuous:3 evening:1 sk:9 additionally:1 nature:3 transfer:1 obtaining:1 improving:1 complex:1 artificially:1 constructing:2 cl:3 did:1 whole:1 arrival:5 profile:1 repeated:1 allowed:1 site:2 representative:1 fashion:1 position:2 wish:3 lie:1 mardia:2 levy:1 third:2 synchronise:1 minute:3 bishop:2 explored:2 incorporating:1 workshop:1 restricting:1 importance:1 occurring:2 demand:4 entropy:2 garcia:2 simply:3 explore:4 likely:3 personnel:1 sport:1 aa:3 utilise:2 acm:2 viewed:1 marked:1 month:5 towards:1 price:1 change:2 typical:3 uniformly:6 total:1 player:1 rarely:1 select:3 crawler:1 outstanding:1 incorporate:2 evaluate:3 tested:2 phenomenon:1 handling:1
1,111
2,013
Grouping with Bias Stella X. Yu Robotics Institute Carnegie Mellon University Center for the Neural Basis of Cognition Pittsburgh, PA 15213-3890 Jianbo Shi Robotics Institute Carnegie Mellon University 5000 Forbes Ave Pittsburgh, PA 15213-3890 stella. yu@es. emu. edu jshi@es.emu.edu Abstract With the optimization of pattern discrimination as a goal, graph partitioning approaches often lack the capability to integrate prior knowledge to guide grouping. In this paper, we consider priors from unitary generative models, partially labeled data and spatial attention. These priors are modelled as constraints in the solution space. By imposing uniformity condition on the constraints, we restrict the feasible space to one of smooth solutions. A subspace projection method is developed to solve this constrained eigenproblema We demonstrate that simple priors can greatly improve image segmentation results. 1 " Introduction Grouping is often thought of as the process of finding intrinsic clusters or group structures within a data set. In image segmentation, it means finding objects or object segments by clustering pixels and segregating them from background. It is often considered a bottom-up process. Although never explicitly stated, higher level of knowledge, such as familiar object shapes, is to be used only in a separate post-processing step. The need for the integration of prior knowledge arises in a number of applications. In computer vision, we would like image segmentation to correspond directly to object segmentation. In data clustering, if users provide a few examples of clusters, we would like a system to adapt the grouping process to achieve the desired properties. In this case, there is an intimate connection to learning classification with partially labeled data. We show in this paper that it is possible to integrate both bottom-up and top-down information in a single grouping process. In the proposed method, the bottom-up grouping process is modelled as a graph partitioning [1, 4, 12, 11, 14, 15] problem, and the top-down knowledge is encoded as constraints on the solution space. Though we consider normalized cuts criteria in particular, similar derivation can be developed for other graph partitioning criteria as well. We show that it leads to a constrained eigenvalue problem, where the global optimal solution can be obtained by eigendecomposition. Our model is expanded in detail in Section 2. Results and conclusions are given in Section 3. 2 Model In graph theoretic methods for grouping, a relational graph GA == (V, E, W) is first constructed based on pairwise similarity between two elements. Associated with the graph edge between vertex i and j is weight Wij , characterizing their likelihood of belonging in the same group. For image segmentation, pixels are taken as graph nodes, and pairwise pixel similarity can be evaluated based on a number of low level grouping cues. Fig. Ic shows one possible defini~ion, where the weight b.etween two pixels is inversely proportional to the magnitude of the strongest intervening edge [9]. a)Image. d)NCuts. e)Segmentation. Figure 1: Segmentation by graph partitioning. a)200 x 129 image with a few pixels marked( +). b)Edge map extracted using quadrature filters.c)Local affinity fields of marked pixels superimposed together. For every marked pixel, we compute its affinity with its neighbours within a radius of 10. The value is determined by a Gaussian function of the maximum magnitude of edges crossing the straight line connecting the two pixels [9]. When there is a strong edge separating the two, the affinity is low. Darker intensities mean larger values. d)Solution by graph partitioning. It is the second eigenvector from normalized cuts [15] on the affinity matrix. It assigns a value to each pixel. Pixels of similar values belong to the same group. e)Segmentation by thresholding the eigenvector with o. This gives a bipartitioning of the image which corresponds to the best cuts that have maximum within-region coherence and between-region distinction. After an image is transcribed into a graph, image segmentation becomes a vertex partitioning problem. Consider segmenting an image into foreground and background. This corresponds to vertex bipartitioning (VI, V2 ) on graph G, where V = VI U V2 and VI n V2 = 0. A good segmentation seeks a partitioning such that nodes within partitions are tightly connected and nodes across partitions are loosely connected. A number of criteria have been proposed to achieve this goal. For normalized cuts [15], the solution is given by some eigenvector of weight matrix W (Fig. Id). Thresholding on it leads to a discrete segmentation (Fig. Ie). W.hile we will focus on normalized cuts criteria [15], most of the following discussions apply to other criteria as well. 2.1 Biased grouping as constrained optimization Knowledge other than the image itself can greatly change the segmentation we might obtain based on such low level cues. Rather than seeing boundaries between black and white regions, we see objects. The sources of priors we consider in this paper are: unitary generative models (Fig. 2a), which could arise from sensor models in MRF [5], partial grouping (Fig. 2b), which could arise from human computer interaction [8], and spatial attention (Fig. 2c). All of these provide additional, often long-range, binding information for grouping. We model such prior knowledge in the form of constraints on a valid grouping configuration. In particular, we see that all such prior knowledge defines a partial a)Bright foreground. b)Partial grouping. c)Spatial attention. Figure 2: Examples of priors considered in this paper. a)Local constraints from unitary generative models. In this case, pixels of light (dark) intensities are likely to be the foreground(background). This prior knowledge is helpful not only for identifying the tiger as the foreground, but also for perceiving the river as one piece. How can we incorporate these unitary constraints into a- graph that handles only pairwise relationships between pixels? b )Global configuration constraints from partial grouping a priori. In this case, we have manually selected two sets of pixels to be grouped together in foreground (+) and background (JJ.) respectively. They are distributed across the image and often have distinct local features. How can we force them to be in the same group and further bring similar pixels along and push dissimilar pixels apart? c)Global constraints from spatial attention. We move our eyes to the place of most interest and then devote our limited visual processing to it. The complicated scene structures in the periphery can thus be ignored while sparing the parts associated with the object at fovea. How can we use this information to facilitate figural popout in segmentation? grouping solution, indicating which set of pixels should belong to one partition. Let Hz, 1 == 1"" ,n, denote a partial grouping. H t have pixels known to be in Vt , t == 1,2. These sets are derived as follows. Unitary generative models: H l and H 2 contains a set of pixels that satisfy the unitary generative models for foreground and background respectively. For example, in Fig. 2a, H l (H2 ) contains pixels of brightest(darkest) intensities. Partial grouping: Each Hz, 1 == 1, ... ,n, contains a set of pixels that users specify to belong together. The relationships between Hz, 1 > 2 and Vt , t == 1,2 are indefinite. Spatial attention: H l == 0 and H 2 contains pixels randomly selected outside the visual fovea, since we want to maintain maximum discrimination at the -fovea but merging pixels far away from the fovea to be one group. To formulate these constraints induced on the graph partitioning, we introduce binary group indicators X == [Xl, X 2 ]. Let N == IVI be the number of nodes in the graph. For t == 1,2, X t is an N x 1 vector where Xt(k) == 1 if vertex k E Vt and 0 otherwise. The constrained grouping problem can be formally written as: min s.t. ?(Xl ,X2 ) Xt(i) == Xt(j), i, j E HE, 1 == 1"" ,n, t == 1,2, Xt(i) =1= Xt(j), i E H l , j E H 2 , t == 1,2, where ?(X1 ,X2 ) is some graph partitioning cost function, such as minimum cuts [6], average cuts [7], or normalized cuts [15]. The first set of constraints can be re-written in matrix form: U T X == 0 , where, e.g. for some column k, Uik == 1, Ujk == ~1. We search for the optimal solution only in the feasible set determined by all the constraints. 2.2 Conditions on grouping constraints The above formulation can be implemented by the maximum-flow algorithm for minimum cuts criteria [6, 13, 3], where two special nodes called source and sink are introduced,.with infinite weights set up between nodes in HI (H2 ) and source(sink). In the context of learning from labeled and unlabeled data, the biased mincuts are linked to minimizing leave-one-out cross validation [2]. In the normalize cuts formulation, this leads to a constrained eigenvalue problem, as soon to be seen. However, simply forcing a few nodes to be in the same group can produce some undesirable graph partitioning results, illustrated in Fig. 3. Without bias, the data points are naturally first organized into top and bottom groups, and then subdivided into left and right halves (Fig. 3a). When we assign points from top and bottom clusters to be together, we do not just want one of the groups to lose its labeled point to the other group (Fig. 3b), but rather we desire the biased grouping process to explore their neighbouring connections and change the organization to left and right division accordingly. Larger Cut a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a 0 a a a a a a a a Desired Cut a a a a a a 0 a 0 a a a a a a a a a a a a a a a a a a a a a a a a a 0 0 0 a a a a a a 0 0 0 0 I 0 a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a .A. a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a Min Cut a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a a 0 a a a a a a a a a a a a a a a a a a a a 0 a a a a a a a a a a a a a a a a a a a a a a II a a a 0 0 a a a a a a a a a a a a a a a a a a a a a a a Perturbed Min Cut b)With bias. a)No bias. Figure 3: Undesired grouping caused by simple grouping constraints. a)Data points are distributed in four groups, with a larger spatial gap between top and bottom groups than that between left and right groups. Defining weights based on proximity, we find the top-bottom grouping as the optimal bisection. b)Introduce two pairs of filled nodes to be together. Each pair has one point from the top and the other from the bottom group. The desired partitioning should now be the left-right division. However, perturbation on the unconstrained optimal cut can lead to a partitioning that satisfies the constraints while producing the smallest cut cost. The desire of propagating partial grouping information on the constrained nodes is, however, not reflected in the constrained partitioning criterion itself. Often, a slightly perturbed version of the optimal unbiased cut becomes the legitimate optimum. One reason for such a solution being undesirable is that some of the "perturbed" nodes-are isolated from their close neighbours. To fix this problem, we introduce the notion of uniformity of a graph partitioning. Intuitively, if two labeled nodes, i and j, have similar connections to their neighbours, we desire a cut to treat them fairly so that if i gets grouped with i's friends, j also gets grouped with j's friends (Fig. 3b). This uniformity condition is one way to propagate prior grouping information from labeled nodes to their neighbours. For normalized cuts criteria, we define the normalized cuts of a single node to be ?.X)- NC u t s ( ~,- EXt(k)=I=Xt(i),YtWik D.. n . This value is high for a node isolated from its close neighbours in partitioning X. We may not know in advance what this value is for the optimal partitioning, but we desire this value to be the same for any pair of nodes preassigned together: NCuts(i;X) == NCuts(j;X), \li,j E Hz, l == 1,??? ,no While this condition does not force NCuts(i; X) to be small for each labeled node, it is unlikely for all of them to have a large value while producing the minimum NCuts for the global bisection. Similar measures can be defined for other criteria. In Fig. 4, we show that the uniformity condition on the bias helps preserving the smoothness of solutions at every labeled point. Such smoothing is necessary especially when partially labeled data are scarce. 0.5 0.5 : -.... .... -.... ~ _o5A~,S/ -0.5 -1 [l...--V_-_l...--_-----'---_----'l o a)Point set data. -1[?..... 300 b)Simple bias. 0 I 100 300 c) Conditioned bias. 0.5 0.5 o o 0 -0.2 -0.4 -0.5 -1 -1 100 d)NCuts wlo bias. e)NCuts 100 wi bias b). f)NCuts wi bias c). Figure 4: Condition constraints with uniformity. a)Data consist of three strips, with 100 points each, numbered from left toright. Two points from the side strips are randomly chosen to be pre-assigned together. b)Simple constraint U T X == 0 forces any feasible solution to have equal valuation on the two marked points. c)Conditioned constraint UTpX == o. Note that now we cannot tell which points are biased. We compute W using Gaussian function of distance with u == 3. d) Segmentation without bias gives three separate groups. e)Segmentation with simple bias not only fails to glue the two side strips into one, but also has two marked points isolated from their neighbours. f)Segmentaion with conditioned bias brings two side strips into one group. See the definition of P below. 2.3 COlllpntation: subspace projection To develop a computational solution for the c9nstrained optimization problem, we introduce some notations. Let the degree matrix D be a diagonal matrix, D ii == Ek Wik, \Ii. Let P == D-IW be the normalized weight matrix. It is a transition probability matrix for nonnegative weight matrix W [10]. Let a == xI~~l be the degree ratio of VI, where 1 is the vector of ones. We define a new variable x == (1 - a)XI - aX2 ? We can show that for normalized cuts, the biased grouping with the uniformity condition is translated into: . mIn E(X) == xT(D-W)x T TD ' s.t. U Px == x x Note, we have dropped the constraint Xt(i) =1= o. Xt(j), i E HI, j E H 2 , t == 1,2. Using Lagrange multipliers, we find that the optimal solution x* satisfies: QPx* == AX*, E(X*) == 1 - A, where Q is a projector onto the feasible solution space: Q == I - D-1V(VTD-1V)-lVT , V == pTU. Here we assume that the conditioned constraint V is of full rank, thus V T D- 1V is invertible. Since 1 is still the trivial 'Solution corresponding to the largest eigenvalue of 1, the second leading right eigenvector of the matrix QP is the solution we seek. To summarize, given weight matrix W, partial grouping in matrix form UT x we do the following to find the optimal bipartitioning: == 0, Compute degree matrix D, D ii == E j Wij , Vi. Compute normalized matrix P == D-1W. Compute conditioned constraint V == pTU. Compute projected weight matrix W == QP==p-n-1V(VTn-1V)-lVTp. Compute the second largest eigenvector x*: Wx* == AX*. Threshold x* to get a discrete segmentation. Step Step Step Step Step Step 1: 2: 3: 4: 5: 6: 3 Results and conclusions We apply our method to the images in Fig. 2. For all the examples, we compute pixel affinity W as in Fig. 1. All the segmentation results are obtained by thresholding the eigenvectors using their mean values. The results without bias, with simple bias U T x == 0 and conditioned bias U T Px == 0 are compared in Fig. 5, 6, 7. e)Simple bias. b) Prior. c)NCuts' on W. f)Seg. on e) g)Conditioned bias. d)Seg. wlo bias. h)Seg. w/ bias. Figure 5: Segmentation with bias from unitary generative models. a)Edge map of the 100 x 150 image. N = 15000. b)We randomly sample 17 brightest pixels for HI (+),48 darkest pixels for H2 (~), producing 63 constraints in total. c) and d) show the solution without bias. It picks up the optimal bisection based on intensity distribution. e) and f) show the solution with simple bias. The labeled nodes have an uneven influence on grouping. g) and h) show the solution with conditioned bias. It successfully breaks the image into tiger and river as our general impression of this image. The computation for the three cases takes 11, 9 and 91ms respectively. Prior knowledge is particularly useful in supplying long-range grouping information which often lacks in data grouping based on low level cues. With our model, the partial grouping prior can be integrated into the bottom-up grouping framework by seeking the optimal solution in a restricted domain. We show that the uniformity constraint is effective in eliminating spurious solutions resulting from simple perturbation on the optimal unbiased solutions. Segmentation from the discretization of the continuous eigenvectors also becomes trivial. e)Simple bias. f)Seg. on e) g)Conditioned bias. h)Seg. w / bias. Figure 6: Segmentation with bias from hand-labeled partial grouping. a)Edge map of the 80 x 82 image. N = 6560. b)Hand-labeled partial grouping includes 21 pixels for HI (+), 31 pixels for H 2 (A), producing 50 constraints in total. c) and d) show the solution without bias. It favors a few largest nearby pieces of similar intensity. e) and f) show the solution with simple bias. Labeled pixels in cluttered contexts are poor at binding their segments together. g) and h) show the solution with conditioned bias. It successfully pops out the pumpkin made of many small intensity patches. The computation for the three cases takes 5, 5 and 71ms respectively. f)4th eig. b) g)6th eig. b) h)4th eig. d) i)6 th eig. d) j)8 th eig. d) Figure 7: Segmentation with bias from spatial attention. N = 25800. a)We randomly choose 86 pixels far away from the fovea (Fig. 2c) for H 2 (A), producing 85 constraints. b) and c) show the solution with simple bias. It is similar to the solution without bias (Fig. 1). d) and e) show the solution with conditioned bias. It ignores the variation in the background scene, which includes not only large pieces of constant intensity, but also many small segments of various intensities. The foreground .successfully clips out the human figure. f) and g) are two subsequent eigenvectors with simple bias. h), i) and j.) are those with conditioned bias. There is still a lot of structural organization in the former, but almost none in the latter. This greatly simplifies the task we face when seeking a segmentation from the continuous eigensolution. The computation for the three cases takes 16, 25 and 220ms respectively. All these benefits come at a computational cost that is 10 times that for the original unbiased grouping problem. We note that we can also impose both UT x == 0 and U T Px == 0, or even U T pBX == 0, S > 1. Little improvement is observed in our examples.' Since projected weight matrix W becomes denser, the computation slows down. We hope that this problem can be alleviated by using multi-scale techniques. It remains open for future research. Acknowledgelllents This research is supported by (DARPA HumanID) ONR NSF IRI-9817496. NOOOI4-00-1-091~ and References [1] A. Amir? and M. Lindenbaum. Quantitative analysis of grouping process. In European Conference on Computer Vision, pages 371-84, 1996. [2] A. Blum and S. Chawla. Learning from labeled and unlabeled data using graph mincuts, 2001. [3] Y. Boykov, O. Veksler, and R. Zabih. Fast approximate energy minimization via graph cuts. In International Conference on Computer Vision, 1999. [4] Y. Gdalyahu, D. Weinshall, and M. Werman. A randpmized algorithm for pairwise clustering. ill Neural Information Processing Systems, pages 424-30, 1998. [5] S. Geman and D. Geman. Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images. IEEE Transactions on Pattern Analysis and Machine Intelligence, 6(6):721-41, 1984. [6] H. Ishikawa and D. Geiger. Segmentation by grouping junctions. In IEEE Conference on Computer Vision and Pattern Recognition, 1998. [7] I. H. Jermyn and H. Ishikawa. Globally optimal regions and boundaries. In International Conference on Computer Vision, 1999. [8] M. Kass, A. Witkin, and D. Terzopoulos. Snakes: Active contour models. International Journal of Computer Vision, pages 321-331, 1988. [9] J. Malik, S. Belongie, T. Leung, and J. Shi. Contour and texture analysis for image segmentation. International Journal of Computer Vision, 2001. [10] M. Meila and J. Shi. Learning segmentation with random walk. ill Neural Information Processing Systems, 2001. [11] P. Perona and W. Freeman. A factorization approach to grouping. In European Conference on Computer Vision, pages 655-70, 1998. [12] J. Puzicha, T. Hofmann, and J. Buhmann. Unsupervised texture segmentation in a deterministic annealing framework. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(8):803-18, 1998. [13] S. Roy and I. J. Cox. A maximum-flow formulation of then-camera stereo correspondence problem. In International Conference on Computer Vision, 1998. [14] E. Sharon, A. Brandt, and R. Basri. Fast multiscale image segmentation. In IEEE Conference on Computer Vision and Pattern Recognition, pages 70-7, 2000. [15] J. Shi and J. Malik. Normalized cuts and image segmentation. In IEEE Conference on Computer Vision and Pattern Recognition, pages 731-7, June 1997.
2013 |@word cox:1 version:1 eliminating:1 glue:1 open:1 seek:2 propagate:1 pick:1 configuration:2 contains:4 pbx:1 segmentaion:1 discretization:1 ka:1 written:2 subsequent:1 partition:3 wx:1 hofmann:1 shape:1 discrimination:2 v_:1 generative:6 cue:3 selected:2 half:1 amir:1 accordingly:1 intelligence:2 supplying:1 node:17 brandt:1 along:1 constructed:1 introduce:4 pairwise:4 multi:1 freeman:1 globally:1 td:1 little:1 becomes:4 notation:1 what:1 weinshall:1 eigenvector:5 developed:2 finding:2 quantitative:1 every:2 jianbo:1 partitioning:16 producing:5 segmenting:1 dropped:1 local:3 treat:1 preassigned:1 ext:1 id:1 vtn:1 might:1 black:1 limited:1 factorization:1 range:2 gdalyahu:1 camera:1 thought:1 projection:2 alleviated:1 pre:1 acknowledgelllents:1 seeing:1 numbered:1 get:3 cannot:1 ga:1 unlabeled:2 onto:1 undesirable:2 close:2 context:2 influence:1 lindenbaum:1 map:3 projector:1 center:1 shi:4 deterministic:1 attention:6 iri:1 cluttered:1 bipartitioning:3 formulate:1 identifying:1 assigns:1 legitimate:1 handle:1 notion:1 variation:1 user:2 neighbouring:1 pa:2 element:1 crossing:1 recognition:3 particularly:1 roy:1 cut:23 geman:2 labeled:14 bottom:9 observed:1 sparing:1 seg:5 region:4 connected:2 defini:1 uniformity:7 segment:3 division:2 basis:1 sink:2 translated:1 darpa:1 various:1 derivation:1 distinct:1 fast:2 effective:1 tell:1 outside:1 encoded:1 larger:3 solve:1 denser:1 otherwise:1 favor:1 itself:2 eigenvalue:3 interaction:1 figural:1 achieve:2 intervening:1 normalize:1 cluster:3 optimum:1 produce:1 leave:1 object:6 help:1 friend:2 develop:1 propagating:1 strong:1 implemented:1 come:1 radius:1 filter:1 stochastic:1 human:2 subdivided:1 assign:1 fix:1 proximity:1 considered:2 ic:1 brightest:2 cognition:1 werman:1 smallest:1 lose:1 iw:1 grouped:3 largest:3 successfully:3 hope:1 minimization:1 sensor:1 gaussian:2 rather:2 derived:1 focus:1 ax:2 june:1 improvement:1 rank:1 likelihood:1 superimposed:1 greatly:3 ave:1 helpful:1 leung:1 unlikely:1 integrated:1 snake:1 spurious:1 perona:1 wij:2 pixel:29 classification:1 ill:2 priori:1 spatial:7 constrained:7 integration:1 special:1 fairly:1 field:1 smoothing:1 never:1 equal:1 manually:1 ishikawa:2 yu:2 unsupervised:1 foreground:7 future:1 few:4 randomly:4 neighbour:6 tightly:1 familiar:1 maintain:1 organization:2 interest:1 light:1 edge:7 partial:11 necessary:1 filled:1 loosely:1 walk:1 desired:3 re:1 isolated:3 column:1 restoration:1 etween:1 cost:3 vertex:4 veksler:1 perturbed:3 eigensolution:1 international:5 river:2 ie:1 invertible:1 together:8 connecting:1 choose:1 transcribed:1 ek:1 leading:1 li:1 includes:2 satisfy:1 explicitly:1 caused:1 vi:5 ax2:1 piece:3 break:1 lot:1 linked:1 capability:1 complicated:1 forbes:1 bright:1 correspond:1 modelled:2 bayesian:1 bisection:3 none:1 straight:1 strongest:1 strip:4 definition:1 energy:1 ptu:2 naturally:1 associated:2 wlo:2 noooi4:1 knowledge:9 ut:2 segmentation:28 organized:1 higher:1 reflected:1 specify:1 formulation:3 evaluated:1 though:1 just:1 hand:2 multiscale:1 eig:5 lack:2 defines:1 brings:1 facilitate:1 normalized:11 unbiased:3 ncuts:9 multiplier:1 former:1 assigned:1 illustrated:1 white:1 undesired:1 criterion:9 m:3 impression:1 theoretic:1 demonstrate:1 bring:1 image:21 boykov:1 jshi:1 vtd:1 qp:2 belong:3 he:1 mellon:2 imposing:1 gibbs:1 smoothness:1 unconstrained:1 meila:1 similarity:2 apart:1 forcing:1 periphery:1 binary:1 onr:1 vt:3 preserving:1 seen:1 additional:1 minimum:3 impose:1 ii:4 full:1 witkin:1 smooth:1 adapt:1 cross:1 long:2 post:1 mrf:1 vision:11 popout:1 robotics:2 ion:1 background:6 want:2 annealing:1 source:3 biased:5 ivi:1 hz:4 induced:1 flow:2 unitary:7 structural:1 ujk:1 restrict:1 simplifies:1 hile:1 stereo:1 jj:1 ignored:1 useful:1 eigenvectors:3 dark:1 clip:1 zabih:1 nsf:1 carnegie:2 discrete:2 group:16 indefinite:1 segregating:1 four:1 threshold:1 blum:1 sharon:1 graph:19 relaxation:1 place:1 almost:1 patch:1 geiger:1 coherence:1 toright:1 hi:4 correspondence:1 nonnegative:1 constraint:24 scene:2 x2:2 nearby:1 min:4 expanded:1 px:3 poor:1 belonging:1 across:2 slightly:1 wi:2 pumpkin:1 intuitively:1 restricted:1 taken:1 remains:1 know:1 junction:1 apply:2 v2:3 away:2 chawla:1 darkest:2 original:1 top:7 clustering:3 especially:1 seeking:2 move:1 malik:2 diagonal:1 devote:1 affinity:5 subspace:2 fovea:5 separate:2 distance:1 separating:1 valuation:1 trivial:2 reason:1 relationship:2 ratio:1 minimizing:1 nc:1 stated:1 slows:1 defining:1 relational:1 perturbation:2 intensity:8 introduced:1 pair:3 connection:3 distinction:1 emu:2 pop:1 below:1 pattern:6 summarize:1 force:3 indicator:1 buhmann:1 scarce:1 wik:1 improve:1 inversely:1 eye:1 stella:2 prior:14 proportional:1 validation:1 eigendecomposition:1 integrate:2 h2:3 degree:3 thresholding:3 supported:1 soon:1 bias:37 guide:1 side:3 terzopoulos:1 institute:2 characterizing:1 face:1 distributed:2 benefit:1 boundary:2 valid:1 transition:1 contour:2 ignores:1 made:1 projected:2 far:2 transaction:2 approximate:1 basri:1 global:4 active:1 pittsburgh:2 belongie:1 xi:2 search:1 continuous:2 european:2 domain:1 arise:2 quadrature:1 x1:1 fig:17 uik:1 darker:1 fails:1 xl:2 intimate:1 down:3 xt:9 grouping:38 intrinsic:1 consist:1 merging:1 texture:2 magnitude:2 conditioned:12 push:1 gap:1 simply:1 likely:1 explore:1 visual:2 lagrange:1 desire:4 partially:3 binding:2 corresponds:2 satisfies:2 extracted:1 goal:2 marked:5 feasible:4 change:2 tiger:2 determined:2 perceiving:1 infinite:1 called:1 mincuts:2 total:2 e:2 indicating:1 formally:1 uneven:1 puzicha:1 latter:1 arises:1 dissimilar:1 incorporate:1
1,112
2,014
A New Discriminative Kernel From Probabilistic Models K. Tsuda,*tM. Kawanabe,* G. Ratsch,?*S . Sonnenburg,* and K.-R. Muller*+ t AIST CBRC, 2-41-6, Aomi, Koto-ku , Tokyo, 135-0064, Japan *Fraunhofer FIRST, Kekulestr. 7, 12489 Berlin , Germany ? Australian National U ni versi ty, Research School for Information Sciences and Engineering, Canberra, ACT 0200 , Australia +University of Pot sdam, Am Neuen Palais 10, 14469 Pot sdam, Germany ko ji. tsuda@aist.go.jp, nabe @first.fraunhofer.de , Gunnar.Raetsch@anu.edu.au , {sonne, klaus }@fir st.fraunhofer.d e Abstract Recently, Jaakkola and Haussler proposed a method for constructing kernel functions from probabilistic models. Their so called "Fisher kernel" has been combined with discriminative classifiers such as SVM and applied successfully in e.g. DNA and protein analysis. Whereas the Fisher kernel (FK) is calculated from the marginal log-likelihood, we propose the TOP kernel derived from Tangent vectors Of Posterior log-odds . Furthermore we develop a theoretical framework on feature extractors from probabilistic models and use it for analyzing FK and TOP. In experiments our new discriminative TOP kernel compares favorably to the Fisher kernel. 1 Introduction In classification tasks , learning enables us to predict the output y E {-1 , + 1} of some unknown system given the input a! E X based on the training examples {a!i ' y;}i=l' The purpose of a feature extractor f : X --+ ]RD is to convert the representation of data without losing the information needed for classification [3] . When X is a vector space like ]Rd , a variety of feature extractors have been proposed (e.g. Chapter 10 in [3]) . However, they are typically not applicable when X is a set of sequences of symbols and does not h ave the structure of a vector space as in DNA or protein analysis [2]. Recently, the Fisher kernel (FK) [6] was proposed to compute features from a probabilistic model p( a!, Y 18). At first, the parameter estimate 9 is obtained from training examples . Then , the tangent vector of the log m arginal likelihood log p( ~ 1 9) is used as a feature vector. The Fisher kernel refers to the inner product in this feature space, but the method is effectively a feature extractor (also since the features are computed explicitly). The Fisher kernel was combined with a discriminative classifier such as SVM and achieved excellent classification result s in several fields, for example in DNA and protein analysis [6 , 5]. Empirically, it is reported that the FK-SVM system often outperforms the classification performance of the plug-in es- timate. 1 Note that the Fisher kernel is only one possible member in the family of feature extractors f iJ (re ) : X --+ ]RD that can be derived from probabilistic models . We call this family "model-dependent feature extractors" . Exploring this family is a very import ant and interesting subject. Since model-dependent feature extractors have been newly developed, the performance measures for them are not yet established. We therefore first propose two performance measures . Then, we define a new kernel (or equivalently a feature extractor) derived from t he Tangent vector Of Posterior log-odds - that we denote as TOP kernel. vVe will analyze the performance of the TOP kernel and the Fisher kernel in terms of our performance measures. Then the TOP kernel is compared favorably to the Fisher kernel in a protein classification experiment. 2 Performance Measures To b egin with, let us describ e the notations. Let re E X b e the input 'point' and y E { -1 , +1 } be the class label. X may be a finite set or an infini te set like ]Rd. Let us assume that we know the parametric model of the joint probability p( re, y19) where 9 E]RP is the parameter vector. Assume that the model p(re,yI9) is regular [7] and contains t he true distribution. Then the true parameter 9 * is uniquely determined. Let iJ be a consistent est imator [1] of 9 *, which is obtained by n training examples drawn i.i.d. from p(re , YI9*). Let oed = of 108i , Vof = (OeJ, ... ,Oep !) T , and V~f denote a p X P matrix whose (i,j) element is 0 2 f 1(08i 08 j ) . As the Fisher kernel is commonly used in combination with linear classifiers such as SVMs, one reasonable performance measure is the classification error of a linear classifier wTfiJ (re) + b (w E]RD and b E]R) in the feature space. Usually wand b are learned, so the optimal feature extractor is different wit h regard to each learning algorithm. To cancel out this ambiguity and to make a theoretical analysis possible, we assume the optimal learning algorithm is used. When wand b are optimally chosen, the classification error is R(fiJ) = min wES ,bE~ E""y<I>[-y(w T fiJ(re ) + b)], (2 .1 ) where S = {w l llwi l = 1,w E ]RD }, <I> [a ] is the step function which is 1 when a > 0 and otherwise 0, and E""y denotes the expectation with respect to the true distribution p( re, y 19*). R(f iJ) is at least as large as the Bayes error L * [3] and R(f iJ) = L * only if the linear classifier implements the same decision rule as the Bayes optimal rule. As a related measure , we consider the estimation error of the posterior probability by a logistic regressor F(w T fiJ(re ) + b), with e.g. F(t) = 1/ (1 + exp( -t)): D(fiJ) = min wE~D ,bE~ E",IF(w T fo(re ) + b) - P(y = +1Ire,9*)I. (2 .2) The relationship between D(fiJ ) and R(fiJ) is illustrated as follows: Let L be t he classification error rate of a posterior probability estimator P(y + lire). With regard t o L, the following inequality is known[l]: L - L* :s; 2E.,IP(y = +l lre ) - P(y = +1 Ire , 9*)I. (2 .3) When P(y +llre):= F(w T fiJ(re) + b) , this inequality leads to the following relationship between the two measures (2.4) R(fiJ) - L* :s; 2D(fiJ)? 1 In classification by plug-in estimate, re is classified by t hresholding the posterior probability fj = sign(P(y = +llre, 0) - 0.5) [1]. --------------------------- Since D(fo ) is an upper bound on R(fo), it is useful to derive a new kernel to minimize D(f 0) ' as will be done in Sec. 4. 3 The Fisher kernel The Fisher kernel (FK) is defined 2 as K (;e , ;e' ) = s(;e ,iJ )TZ-1(iJ)s (;e' ,iJ) , where s is the Fisher score s(;e ,iJ ) = (otl1logp(;eliJ) , ... ,Otlp 10gp( ;eliJ )) T = \7 e logp(;e ,iJ ), and Z is the Fisher information matrix: Z(9) = E", [s(;e,9)s(;e,9)TI9]. The theoretical foundation of FK is described in the following theorem [6]: "a kernel classifier employed the Fisher kernel derived from a model that cont ains the lab el as a la tent variable is , asymptotically, at least as good a classifier as t he MAP labeling based on the model" . The theorem says that the Fisher kernel can perform at least as well as the plug-in estimate, if the parameters of linear classifier are properly det ermined (cf. Appendix A of [6]). With our p erforman ce measure, this t heorem can be represented more concisely: R(f 0) is bounded by the classificat ion error of t he plug-in estimate R(fo) :S; E""y<I> [- y(P(y = + ll;e,iJ ) - 0. 5)] . (3.1 ) Not e that the classification rule constructed b y the plug-in estimate P( y = + 11;e , iJ) can also be realized by a linear classifier in feat ure space. Propert y (3.1) is important since it gu arantees that the Fisher kernel performs better when t he optimal w and b are available. However, the Fisher kernel is not the only one to satisfy t his inequality. In the next section, we present a new kernel which satisfies (3.1) and has a more app ealing theoretical property as well. 4 The TOP Kernel Definition Now we proceed to propose a new kernel: Our aim is to obtain a feature extractor that achieves small D(f 0). When a feature extractor!0 (;e) satisfies3 W T !o(;e ) + b = p -1( p(y = + 11;e , 9 *)) for all;e E X (4 .1 ) with certain values of w and b, we have D(f 0) = O. However, since the true parameter 9 * is unknown, all we can do is t o construct ! 0 which approximately satisfies (4.1). Let us define v( ;e,9) = p-1 (p (y = +11;e , 9 )) = 10g( P( y = +11;e,9 )) -log(P (y = -11;e,9) ), which is called the posterior log-odds of a probabilistic model [1]. By Taylor expansion around the estimate iJ up to t he first order 4 , we can approximate v( ;e,9*) as l' v( ;e,9*) ~ v( ;e,iJ) + L0tliv( ;e ,iJ)(e: - ad. (4.2) i=l 2In practice, some variants of the Fisher kernel are used. For example, if the derivative of each class distribution , not marginal , is taken, the feature vector of FK is quite simila r to that of our kernel. However , th ese variants should b e deliberately discrimin at ed from the Fisher kernel in theoretical disc ussions. Throughout this pap er including ex p erim ents, we adopt the o rigi nal defi ni t ion of the F isher kern el from [6] . 3Notice t h at p- l (t) = log t - log(l - t ) 40 ne can easily derive TOP kern els from higher order Taylor ex pansions . Howeve r, we will only deal wit h t he first order expansion here, because higher order ex pansio ns would induce extremely high dimensional feature vectors in practical cases. Thus , by setting ( 4.3) and w:= w * = (1, 8; - e l , ??? , 8; - ep)T, b = 0, (4.4) equation (4.1) is approximately satisfied. Since a Tangent vector Of the Posterior log-odds const itutes the main p art of the feature vector, we call the inner product of t he two feature vectors "TOP kernel" : (4.5) It is easy to verify t hat the TOP kernel satisfies (3.1) , b ecause we can construct the same decision rule as the plug-in estimate by using the first element only (i.e. w = (1 , 0, . .. ,0), b = 0). A Theoretical Analysis In this section , we compare the TOP kernel with the plugin es timate in terms of p erformance measures . Later on , we assume that 0 < P (y = +1Ial,8 ) < 1 to prevent IV( al,8)1 from going to infinity. Also, it is assumed t hat VeP (y = +1Ial , 8) and V~P ( y = +1Ial,8 ) are bounded. Substituting the plug-in estimate denoted by the subscript IT into D(fo ), we have Define 68 = 8- 8*. By Taylor expansion around 8* , we have where 8 0 = 8* + "(68 (O :S "( :S 1). When the TOP kernel is used, D(fo) :S E",IF((w* )T fo(al)) - P (y = +1 Ial,8*)I , (4.7) where w * is defined as in (4.4). Since P is Lipschitz-continuous, there is a finit e positive constant M such that IP(a) - P (b)1 :S Mia - bl. Thus, D(fo) :S ME",I(w *)T fo (rn ) - P-l (P (y = + 1Irn , 8* )) I? (4.8) Since (w* )T f 0 (al ) is the T aylor expansion of p - 1 (P(y = + 11al , 8 *)) up to the first order (4.2) , the first order t erms of 68 are excluded from the right side of (4.8 ), thus D(fo ) = 0 (11 68 112 ) . Since both, the plug-in and the TOP kernel, dep end on the parameter estimate 8, the errors D,,(8) and D(fo) become smaller as 1168 11 decreases. This shows t h at if w and b are optimally chosen , t he rate of convergence of the TOP kernel is much faster than that of the plug-in estimate. This result is closely related to large sample p erforman ces : Assuming t hat 8 is a n 1/ 2 consistent estimator with asymptotic normality (e.g. the maximum likelihood estimator) , we have 11681 1 = Op(n- l / 2 )[7J, where 01' denotes stochastic order cf. [1]. So we can directly derive the convergen ce order as D,,(8) = Op (n- l / 2 ) and D(f 0) = Op( n - l ). By using t he rela tion (2.4) , it follows that R,, (8 ) - L* = Op( n - l / 2 ) and R(f 0) - L * = Op (n- l ).5 Therefore, t he TOP kernel h as a much b etter convergen ce rate in R(f 0)' which is a strong motiva tion to use the TOP kernel instead of plug-in estimate. 5Fo r detail ed disc ussion a bout t he conve rgence orders of classificatio n e rror, see C ha pte r 6 of [1] However, we must notice that this fast rate is possible only when the optimal linear classifier is combined with the TOP kernel. Since non-optimal linear classifiers typically have the rate Op(n- 1 / 2 ) [1 ], the overall rate is dominated by the slower rate and turns out to be Op (n - 1 / 2 ) . But this theoretical analysis is still meaningful, because it shows the existence of a very efficient linear boundary in the TOP feature space. This result encourages practical efforts to improve linear boundaries by engineering loss functions and regularization terms with e.g. cross validation, bootstrapping or other model selection criteria [1]. Exponential Family: A Special Case ?When the distribution of two classes belong to the exponential family, the TOP kernel can achieve an even better result than shown above . Distributions of the exponential family can be written as q( re , 11) = exp( 11 T t (;I!) +~( 11)) , where t (;I!) is a vector-valued function called sufficient statistics and ~ ( 11) is a normalization factor [4]. Let 0' denote the parameter for class prior probability of the positive model P( y = +1). Then, the probabilistic model IS described as where 8 = {O' , 11+1 ' 11 - 1}? The posterior log-odds reads The TOP feature vector is described as A A AT ATT iiJ(;I! ) = (v( ;I! ,8) ,Oav(re ,8 ), V'7 +1 v(re , 8 ) , V'7 _1 v (;I!,8 ) ) . where V'7 ,v(;I!,iJ ) = s(ts(re) + V'7,~s(f,s)) for s = {+1,-1}. So, when w = ( 1,0, 11+1 - "'+1,11"-1 - "'- 1) T and b is properly set, the true log-odds p - l (P(y = +11;I!,8*)) can be constructed as a linear function in the feature space (4.1). Thus DUiJ) = 0 and RUiJ) = L*. Furthermore, since each feature is represented as a linear function of sufficient statistics t+1 (re) and t - l (re), one can construct an equivalent feature space as (t + l (re) T, Ll (re) T) T without knowing iJ. This result is important because all graphical models without hidden states can be represented as members of the exponential family, for example markov models [4] . 5 Experiments on Protein Data In order to illustrate that the TOP kernel works well for real-world problems , we will show t he result s on protein classification. The protein sequence data is obtained from the Superfamily websit e. 6 This site provides sequence files with different degrees of redundancy filtering ; we used the one with 10% redundancy filtering. Here, 4541 sequences are hierarchically labeled into 7 classes, 558 folds, 845 superfamilies and 1343 families according to the SCOP(1.53) scheme. In our experiment , we determine the top category "classes" as the learning target. The numbers of sequences in the classes are listed as 791, 1277, 1015 , 915,84,76,383 . We only use the first 4 classes, and 6 two-class problem s are generated from all pairs among t he 4 classes . The 5th and 6th classes are not used because t he number of examples is too small. Also, the 7th class is not used because this class is quite different from the others and too easy to classify. In each two-class problem , the examples are randomly divided into 25 % training set, 25 % validation set and 50% t est set. The validation set is used for model selection. 6http://stash.mrc-lmb.cam.ac.uk/S UPERFAMILY / As a probabilistic model for protein sequences, we make use of hidden markov models [2] with fully connected states. 7 The Baum- Welch algorithm (e.g. [2]) is used for maximum likelihood training. To construct FK and TOP kernels , the derivatives with respect to all paramet ers of the HMMs from both classes are included. The derivative with respect to the class prior probability is included as well: Let q( OIl , e) be the probability density function of a HMM. Then, the marginal di stribution is written as p(ocI8) = aq( oc, e+1 ) + (1- a)q( oc, L1) , where a is a parameter corresponding to the class prior. The feature vector of FK consists of the following: V'e, logp( oc I8) 00: logp(oc I8) P (y=s loc , 8)V'e , logq(oc ,es) 1 SE {-l ,+l } , 1 ' --;;-P (y = +1 1?c , 9) - - -, P(y = -11?c, 9) , a I -a (5.1 ) (5.2) while the feature vector of TOP includes V'e ,v( oc ,8) sV'e , logq( oc ,e s) s = {+ 1, _ 1}.8 We obtained e+1 and e-1 from the training examples of respective classes and set a = 0.5. In the definition of the TOP kernel (4.5), we did not include any normalization of feature vectors. However, in practical situations, it is effective to normalize feature s for improving classification performance. Here, each feature of the TOP kernel is normalized to have mean 0 and variance 1. Also in FK, we normalized the features in the same way instead of using the Fisher information matrix, because it is difficult to estimate it reliably in a high dimensional parameter space. Both, the TOP kernel and FK are combined with SVMs with bias terms. When classifying with HMMs , one observes the difference of the log-likelihoods for the two classes and discriminat es by thresholding at an appropriate value. Theoretically, this threshold should be determined by the (true) class prior probability. But, this is typically not available. Furthermore the estimation of the prior probability from training data often leads to poor results [2] . To avoid this problem, the threshold is determined such that the false positive rate and the false negative rate are equal in the test set. This threshold is determined in the same way for FK-SVMs and TOP-SVMs. The hybrid HMM-TOP-SVM system has several model parameters: the number of HMM states, the pseudo count value [2] and the regularization parameter C of the SVM. vVe determine these parameters as follows: First, the number of states and the pseudo count value are determined such that the error of the HMM on the validation set (i. e. validation error) is minimized. Based on the chosen HMM model, the paramet er C is det ermined such that the validation error of TOP-SVM is minimized. Here, the number of states and the pseudo count value are chosen from {3 , 5,7,10,15,20,30,40, 60} and {l0 -1 0, 10 - 7 , 10 - 5 , 10 - 4 ,10 - 3 , 1O- 2 }, respectively. For C, 15 equally spaced points on the log scale are taken from [10- 4 ,10 1]. Note that the model selection is performed in the same manner for the Fisher kernel as well. The error rates over 15 different training/validation/test divisions are shown in Figure 1 and 2. The results of stat istical tests are shown in Table 1 as well. Compared with the plug-in est imate, the Fisher kernel performed significantly better in several sett ings (i.e. 1-3, 2-3, 3-4). This result partially agrees with observations in [6]. However, our TOP approach significantly outperforms the Fisher kernel: According to the Wilcoxson signed ranks test, the TOP kernel was significantly better 7Several HMM models have been engineered for protein classification [2]. However, we do not use such HMMs because the main purpose of experiment is to compare FK and TOP. 8 0aV (OC, 0) is a constant which does not depend on OIl. So it is not included in the feature vector. 1-2 1-3 0.3 0.1 6 ~ 0.1 4 0.1 2 0.1 0.08 0.25 0.2 ~ FK TOP 2-3 0.18 0.16 0.14 ~ P ""I 0. 32 0.3 0. 28 1 0.26 FK TOP P 2-4 0.32 FK TOP 3-4 04 0.3 t8 ~ 0.12 0.1 P ~$ ~~~ 1 0. 24 0.15 P 1-4 0. 36 FK 0.28 0.26 0.24 ffi 0.22 0. 2 TOP !~ ,~ ~ 1 :::1 034 032 , 03 0 28 1 I P FK TOP P FK TOP Figure 1: The error rates of SVMs with two feature extractors in t he protein classification experiments. T he labels 'P ','FK' ,'T OP' denote t he p lug-in estimate , the F isher kernel and t he TOP kernel, respect ively. T he t itle on each subfigure shows the t wo prot ein classes used for t he experiment. 1-2 F igure 2: Comparison of the error rates of t he F isher kernel and t he TO P kernel in discrim ination between class 1 and 2. Every point corresponds to one of 15 differen t t raining/validation /test set splits. Except t wo cases, t he T OP kernel achieves smaller error rates . 0.14 0.12 0.1 0.08 0?ct~06 0.08 0.1 0.12 0.14 TOP in all settings . Also, the t-test j udged t h at t he difference is significant except for 1-4 and 2-4. This indicates that the T OP kernel was able to capture discrim inative information better than t he Fisher kernel. 6 Conclusion In this study, we presented the new discrim inative TOP kernel derived from probabilistic models. Since the theoret ical framework for su ch kernels has so far not been established, we proposed two performance measures to analyze them and gave bounds an d rates to gain a bett er insigh t into model depen dent feat ure extractors from probabilistic models. Exp erimentally, we showed that the T OP kernel compares favorably to F K in a realistic protein classification experim ent . Note t h at Sm ith and Gales[8] h ave sh own t h at a similar approach works excellent ly in speech recogni tion tasks as well. Fu t ure research will focus on constructing sm all sam ple bounds fo r t he T OP kernel to exten d the validity of t his work. Since other nonlinear transformat ions F are possible for obtaining different and possibly b etter features, we will furt hermore consider to learn t he nonlinear transformat ion F from training samples . An interes ting point is that so far T OP kernels perform local linear approximations, it would be interesting to move in the direction of local or even Table 1: P-values of statistical test s in the protein classification experiments . Two kinds of tests, t- test (denoted as T in the table) and Wilcoxson signed ranks test (i.e. WX) , are used. When the difference is significant (p-value < 0.05), a single star * is put beside the value. Double stars ** indicate that the difference is very significant (p-value < 0.01). I Methods I Test II P, FK P, TOP FK,TOP T WX T WX T WX I Methods I Test II P, FK P, TOP FK,TOP T WX T WX T WX 1-2 0.95 0.85 0.015* 4.3 X 10- 4 ** 0.0093** 8.5 X 10 - 4 ** 1-3 0.14 0.041 * 1.7 X 10 - ~** 6.1 X 10- 5 ** 2.2 X 10 -4** 6.1 X 10 - 5 ** 1-4 0.78 0.24 0.11 0.030* 0.21 0.048* 2-3 0.0032** 0.0040** 3.0 X 10 -1~** 6.1 X 10 - 5 ** 2.6 X 10 -M* 6.1 X 10 - 5 ** 2-4 0.79 0.80 0.059 0.035* 0.079 0.0034** 3-4 0.12 0.026* 5.3 X 10 -0** 3.1 X 10- 4 ** 0.0031 ** 1.8 X 10- 4 ** global nonlinear expansions. Acknowledgment s vVe thank T. Tanaka, M . Sugiyama, S.-I. Amari, K. Karplus, R. Karchin, F. Sohler and A. Zien for valuable discussions. Moreover, we gratefully acknowledge partial support from DFG (JA 379/9-1, MU 987/1-1) and travel grants from EU (Neurocolt II). References [1] L. Devroye, L. Gyorfi, and G. Lugosi. A Probabilistic Theory of Pattern Recognition. Springer , 1996. [2] R. Durbin, S. Eddy, A. Krogh, and G. Mitchison. Biological Sequence Analysis: Probabilistic Models of Proteins and Nucleic Acids. Cambridge University Press , 1998. [3] K. Fukunaga. Introduction to Statistical Pattern Recognition. Academic Press, San Diego, 2nd edition, 1990. [4] D. Geiger and C. Meek. Graphical models and exponential famili es . Technical Report MSR- TR-98-10, Microsoft Research, 1998. [5] T.S. Jaakkola, M. Diekhans, and D. Haussler. A discriminative framework for detecting remote protein homologies. J. Compo Biol. , 7:95-114, 2000. [6] T.S. Jaakkola and D. Haussler. Exploiting generative models in discriminative classifiers. In M.S. Kearns, S.A. SoHa, and D.A. Cohn, editors, Advances in Neural Information Processing Systems 11 , pages 487- 493. MIT Press, 1999. [7] N. Murata, S. Yoshizawa, and S. Amari. Network information criterion determining the number of hidden units for an artificial neural network model. IEEE Trans. Neural Networks, 5:865- 872, 1994. [8] N. Smith and M. Gales. Speech recognition using SVMs. In T.G. Dietterich, S. Becker, and Z. Ghahramani , editors, Advances in Neural Information Processing Systems 14. MIT Press, 2002. to appear.
2014 |@word msr:1 nd:1 tr:1 contains:1 score:1 att:1 loc:1 outperforms:2 erms:1 yet:1 import:1 must:1 written:2 realistic:1 wx:7 enables:1 generative:1 ial:4 ith:1 smith:1 compo:1 ire:2 provides:1 detecting:1 istical:1 constructed:2 become:1 consists:1 manner:1 theoretically:1 hresholding:1 notation:1 bounded:2 moreover:1 kind:1 developed:1 bootstrapping:1 pseudo:3 every:1 act:1 classifier:12 prot:1 yi9:2 uk:1 unit:1 ly:1 grant:1 appear:1 positive:3 engineering:2 local:2 plugin:1 analyzing:1 ure:3 subscript:1 approximately:2 lugosi:1 signed:2 au:1 hmms:3 gyorfi:1 practical:3 acknowledgment:1 practice:1 implement:1 erformance:1 lira:1 significantly:3 induce:1 refers:1 regular:1 kern:2 protein:14 sonne:1 selection:3 put:1 equivalent:1 map:1 baum:1 go:1 welch:1 wit:2 rule:4 haussler:3 estimator:3 his:2 target:1 diego:1 losing:1 element:2 defi:1 recognition:3 labeled:1 logq:2 ep:1 capture:1 connected:1 sonnenburg:1 eu:1 decrease:1 remote:1 observes:1 valuable:1 mu:1 cam:1 oed:1 ings:1 depend:1 division:1 gu:1 finit:1 easily:1 joint:1 chapter:1 represented:3 fast:1 effective:1 artificial:1 labeling:1 klaus:1 whose:1 quite:2 valued:1 say:1 otherwise:1 amari:2 statistic:2 gp:1 ip:2 sequence:7 propose:3 product:2 achieve:1 normalize:1 ent:1 exploiting:1 convergence:1 lmb:1 double:1 derive:3 develop:1 transformat:2 stat:1 illustrate:1 ac:1 ij:16 op:13 school:1 dep:1 strong:1 krogh:1 pot:2 indicate:1 australian:1 direction:1 fij:9 closely:1 tokyo:1 koto:1 stochastic:1 australia:1 engineered:1 ja:1 discriminat:1 biological:1 paramet:2 exploring:1 dent:1 around:2 exp:3 predict:1 heorem:1 substituting:1 achieves:2 adopt:1 purpose:2 estimation:2 travel:1 applicable:1 label:2 agrees:1 successfully:1 mit:2 aim:1 avoid:1 jaakkola:3 derived:5 l0:1 focus:1 properly:2 rank:2 likelihood:5 indicates:1 ave:2 am:1 dependent:2 el:3 typically:3 hidden:3 irn:1 ical:1 going:1 germany:2 overall:1 classification:16 among:1 denoted:2 art:1 special:1 neuen:1 marginal:3 field:1 construct:4 equal:1 sohler:1 cancel:1 minimized:2 others:1 report:1 randomly:1 duij:1 national:1 dfg:1 ents:1 microsoft:1 sh:1 furt:1 ively:1 fu:1 partial:1 respective:1 iv:1 taylor:3 re:20 tsuda:2 karplus:1 theoretical:7 subfigure:1 classify:1 versi:1 logp:3 too:2 optimally:2 reported:1 sv:1 combined:4 st:1 density:1 infini:1 probabilistic:12 regressor:1 aomi:1 ambiguity:1 satisfied:1 lug:1 gale:2 exten:1 fir:1 possibly:1 tz:1 derivative:3 japan:1 de:1 scop:1 star:2 sec:1 includes:1 satisfy:1 explicitly:1 ad:1 later:1 tion:3 performed:2 lab:1 analyze:2 bayes:2 minimize:1 kekulestr:1 ni:2 variance:1 acid:1 murata:1 spaced:1 ant:1 disc:2 mrc:1 app:1 classified:1 mia:1 fo:13 ed:2 definition:2 ty:1 yoshizawa:1 di:1 timate:2 gain:1 newly:1 eddy:1 recogni:1 higher:2 done:1 furthermore:3 su:1 nonlinear:3 cohn:1 logistic:1 convergen:2 oil:2 dietterich:1 validity:1 verify:1 true:6 homology:1 deliberately:1 normalized:2 regularization:2 excluded:1 read:1 aist:2 illustrated:1 deal:1 llwi:1 ll:2 uniquely:1 encourages:1 oc:8 criterion:2 palais:1 performs:1 l1:1 fj:1 recently:2 ji:1 empirically:1 lre:1 jp:1 belong:1 he:29 raetsch:1 significant:3 cambridge:1 rd:6 fk:24 vep:1 depen:1 sugiyama:1 gratefully:1 aq:1 pte:1 posterior:8 own:1 showed:1 certain:1 inequality:3 muller:1 employed:1 determine:2 ii:3 zien:1 technical:1 faster:1 academic:1 plug:11 cross:1 divided:1 equally:1 variant:2 ko:1 expectation:1 kernel:63 normalization:2 achieved:1 ion:4 cbrc:1 whereas:1 ratsch:1 file:1 subject:1 member:2 odds:6 call:2 split:1 easy:2 variety:1 discrim:3 gave:1 nabe:1 inner:2 tm:1 knowing:1 det:2 diekhans:1 ese:1 becker:1 effort:1 wo:2 speech:2 proceed:1 useful:1 se:1 listed:1 svms:6 category:1 dna:3 wes:1 http:1 bett:1 notice:2 sign:1 gunnar:1 redundancy:2 threshold:3 drawn:1 prevent:1 ce:4 nal:1 asymptotically:1 pap:1 convert:1 wand:2 family:8 reasonable:1 throughout:1 geiger:1 decision:2 appendix:1 bound:3 ct:1 meek:1 igure:1 fold:1 ti9:1 durbin:1 infinity:1 propert:1 dominated:1 min:2 extremely:1 fukunaga:1 according:2 combination:1 poor:1 smaller:2 sam:1 tent:1 taken:2 equation:1 imate:1 turn:1 count:3 ffi:1 needed:1 know:1 end:1 available:2 etter:2 experim:1 kawanabe:1 appropriate:1 differen:1 slower:1 rp:1 hat:3 existence:1 top:46 denotes:2 cf:2 include:1 graphical:2 const:1 ting:1 ghahramani:1 bl:1 move:1 realized:1 parametric:1 thank:1 berlin:1 neurocolt:1 hmm:6 me:1 assuming:1 devroye:1 cont:1 relationship:2 equivalently:1 difficult:1 favorably:3 negative:1 reliably:1 unknown:2 perform:2 upper:1 av:1 observation:1 nucleic:1 markov:2 sm:2 finite:1 acknowledge:1 t:1 situation:1 rn:1 pair:1 concisely:1 learned:1 bout:1 established:2 tanaka:1 trans:1 able:1 usually:1 pattern:2 including:1 hybrid:1 normality:1 scheme:1 improve:1 ne:1 rela:1 fraunhofer:3 prior:5 tangent:4 determining:1 asymptotic:1 beside:1 loss:1 fully:1 interesting:2 filtering:2 conve:1 validation:8 foundation:1 degree:1 sufficient:2 consistent:2 thresholding:1 editor:2 i8:2 classifying:1 side:1 bias:1 superfamily:2 regard:2 boundary:2 calculated:1 raining:1 world:1 commonly:1 san:1 ple:1 far:2 approximate:1 feat:2 global:1 inative:2 assumed:1 discriminative:6 mitchison:1 continuous:1 table:3 ku:1 learn:1 obtaining:1 improving:1 expansion:5 excellent:2 constructing:2 t8:1 did:1 main:2 hierarchically:1 edition:1 vve:3 site:1 canberra:1 theoret:1 ein:1 n:1 iij:1 exponential:5 extractor:13 theorem:2 er:4 symbol:1 svm:6 false:2 effectively:1 te:1 anu:1 soha:1 partially:1 rror:1 springer:1 ch:1 corresponds:1 satisfies:3 classificat:1 lipschitz:1 fisher:25 included:3 determined:5 except:2 kearns:1 called:3 e:5 la:1 est:3 meaningful:1 support:1 biol:1 ex:3
1,113
2,015
Geometrical Singularities in the Neuromanifold of Multilayer Perceptrons Shun-ichi Amari, Hyeyoung Park, and Tomoko Ozeki Brain Science Institute, RIKEN Hirosawa 2-1, Wako, Saitama, 351-0198, Japan {amari, hypark, tomoko} @brain.riken.go.jp Abstract Singularities are ubiquitous in the parameter space of hierarchical models such as multilayer perceptrons. At singularities, the Fisher information matrix degenerates, and the Cramer-Rao paradigm does no more hold, implying that the classical model selection theory such as AIC and MDL cannot be applied. It is important to study the relation between the generalization error and the training error at singularities. The present paper demonstrates a method of analyzing these errors both for the maximum likelihood estimator and the Bayesian predictive distribution in terms of Gaussian random fields, by using simple models. 1 Introduction A neural network is specified by a number of parameters which are synaptic weights and biases. Learning takes place by modifying these parameters from observed input-output examples. Let us denote these parameters by a vector () = (0 1 , .. . , On). Then, a network is represented by a point in the parameter space S, where () plays the role of a coordinate system. The parameter space S is called a neuromanifold. A learning process is represented by a trajectory in the neuromanifold. The dynamical behavior of learning is known to be very slow, because of the plateau phenomenon. The statistical physical method [1] has made it clear that plateaus are ubiquitous in a large-scale perceptron. In order to improve the dynamics of learning, the natural gradient learning method has been introduced by taking the Riemannian geometrical structure of the neuromanifold into account [2, 3]. Its adaptive version, where the inverse of the Fisher information matrix is estimated adaptively, is shown to have excellent behaviors by computer simulations [4, 5]. Because of the symmetry in the architecture of the multilayer perceptrons, the parameter space of the MLP admits an equivalence relation [6, 7]. The residue class divided by the equivalence relation gives rise to singularities in the neuromanifold, and plateaus exist at such singularities [8]. The Fisher information matrix becomes singular at singularities, so that the neuromanifold is strongly curved like the spacetime including black holes. In the neighborhood of singularit ies, the Fisher-Cramer-Rao paradigm does not hold, and the estimator is no more subject to the Gaussian distribution even asymptotically. This is essential in neural learning and model selection. The AlC and MDL criteria of model selection use the Gaussian paradigm, so that it is not appropriate. The problem was first pointed out by Hagiwara et al. [9]. Watanabe [10] applied algebraic geometry to elucidate the behavior of the Bayesian predictive estimator in MLP, showing sharp difference in regular cases and singular cases. Fukumizu [11] gives a general analysis of the maximum likelihood estimators in singular statistical models including the multilayer perceptrons. The present paper is a first step to elucidate effects of singularities in the neuromanifold of multilayer perceptrons. We use a simple cone model to elucidate how different the behaviors of the maximum likelihood estimator and the Bayes predictive distribution are from the regular case. To this end, we introduce the Gaussian random field [11, 12, 13], and analyze the generalization error and training error for both the mle (maximum likelihood estimator) and the Bayes estimator. 2 Topology of neuromanifold Let us consider MLP with h hidden units and one output unit, h Y= L Vi<{J (Wi? x) + n. (1) i= l where y is output, x is input and n is Gaussian noise. Let us summarize all the parameters in a single parameter vector () = (Wl , ???, Wh; Vl , ???, Vh) and write h f(x; ()) = L Vi<{J (Wi? x). (2) i=l Then, () is a coordinate system of the neuromanifold. Because of the noise, the input-output relation is stochastic, given by the conditional probability distribution p(ylx,()) = 1 {I -2(y-f(x;())) 2} , J2 exp (3) where we normalized the scale of noise equal to 1. Each point in the neuromanifold represents a neural network or its probability distribution. It is known that the behavior of MLP is invariant under 1) permutations of hidden units , and 2) sign change of both Wi and Vi at the same time. Two networks are equivalent when they are mapped by any of the above operations which form a group. Hence, it is natural to treat the residual space SI ::::J, where ::::J is the equivalence relation. There are some points which are invariant under a some nontrivial isotropy subgroup, on which singularities occurs. When Vi = 0, vi<{J (Wi? x) = 0 so that all the points on the sub manifold Vi = 0 are equivalent whatever Wi is. We do not need this hidden unit. Hence, in M = SI ::::J, all of these points are reduced to one and the same point. When Wi = Wj hold, these two units may be merged into one, and when Vi +Vj is the same, the two points are equivalent even when they differ in Vi - Vj. Hence, the dimension reduction takes place in the subspace satisfying Wi = Wj. Such singularities occur on the critical submanifolds of the two types (4) 3 Simple toy models Given training data, the parameters of the neural network are estimated or trained by learning. It is important to elucidate the effects of singularities on learning or estimation. We use simple toy models to attack this problem. One is a very simple multilayer percept ron having only one hidden unit. The other is a simple cone model: Let x be Gaussian random variable x E R d +2 , with mean p, and identity covariance matrix I , (5) and let 5 = {p,Ip, E R d +2 } be the parameter space. The cone model M is a subset of 5, embedded as M : p, (6) where c is a constant, IIa 2 11 = 1, When d = 1, 51 is a circle so that p, = W W E 5 d and 5 d is a d-dimensional unit sphere. is replaced by angle B, and we have ~ VI + c2 See Figure 1. The M is a cone, having is the singular point. (~, 1 B) . ( ccos (7) csinB w) as coordinates, where the apex ~ = 0 ,, Figure 1: One-dimensional cone model The input-output relation of a simple multilayer perceptron is given by y = v<p(w . x) +n (8) When v = 0, the behavior is the same whatever w is. Let us put w = (3w , where (3 = Iwl and W E 5 d , and ~ = vlwl, 'l/J( x;(3 , w) = <p{(3(w? x)} /(3. We then have y = ~'l/J(x;(3,w) +n (9) which shows the cone structure with apex at ~ = O. In this paper, we assume that (3 is knwon and does not need to be estimateed. 4 Asymptotic statistical inference: generalization error and training error Let D = {Xl,???, XT} be T independent observations from the true distribution Po(x) which is specified by ~ = 0, that is, at the singular point. In the case of neural networks , the training set D is T input-output pairs (Xt, Yt), from the conditional probability distributions p(Ylx;~, w) and the true one is at ~ = O. The discussions go in parallel, so that we show here only the cone model. We study the characteristics of both the mle and the Bayesian predictive estimator. Let p(x) be the estimated distribution from data D . In the case of mle, it is given by p(x; 0) where 0 is the mle given by the maximizer of the log likelihood. For the Bayes estimator, it is given by the Bayes predictive distribution p(xID). We evaluate the estimator by the generalization error defined by the KL-divergence from Po(x) to p(x), Eg en = ED [K[po : pll, K[Po: p] = Epo [log ~(~i] . (10) Similarly, the training error is defined by using the empirical expectation, (11) In order to evaluate the estimator p, one uses E gen , but it is not computable. Instead, one uses the Etrain which is computable. Hence, it is important to see the difference between Egen and Etrain- This is used as a principle of model selection. When the statistical model M is regular, or the true distribution Po (x) is at a regular point , the mle-based p(x, 0) and the Bayes predictive distribution are asymptotically equivalent, and are Fisher efficient under reasonable regularity conditions, Eg en ~ d 2T ' Eg en ~ Etrain d + T' where d is the dimension number of parameter vector (12) (j. All of these good relations do not hold in the singular case. The mle is no more asymptotically Gaussian, the mle and the Bayes estimators have different asymptotic characteristics, although liT consistency is guaranteed. The relation between the generalization and training error is different, so that we need a different model selection criterion to determine the number of hidden units. 5 Gaussian random fields and mle Here, we introduce the Gaussian random field [11, 12, 13] in the case of the cone model. The log likelihood of data D is written as L(D,~,w) = 1 T -"2l: Ilxt - ~a(w)112. (13) t=l Following Hartigan [13] (see also [11] for details), we first fix wand search for the ~ that maximizes L. This is easy since L is a quadratic function of ( The maximum t is given by (14) = Y(w) (15) a(w) ? X, By the central limit theorem, Y (w) = a( w) . x is a Gaussian random field defined on Sd = {w}. By substituting t(w) in (14) the log likelihood function becomes T , I", 1 2 -2 ~ IIXtl1 + 2Y L(w) = 2 (w). (16) t=l Therefore, the mle Theorem 1. w is given by the maximizer of L(w), w= argmaxwy2(w). In the case of the cone model, the mle satisfies 2~ED h~p y Egen Etrain Corollary 1. = - 2~ED 2 h:p y (w)] , 2 (w)] . (17) (18) When d is large, the mle satisfies Egen :::::: Etrain :::::: c2 d 2T(1 + c2 ) ' c2 d 2T(1 + c2 )? (19) (20) It should be remarked that the generalization and training errors depend on the shape parameter c as well as the dimension number d. 6 Bayesian predictive distribution The Bayes paradigm uses the posterior probability of the parameters based on the set of observations D. The posterior probability density is written as, T p(~,wID) = c(D)1f(~,w) rrp(xtl~, w) , (21) t= l where c(D) is the normalization factor depending only on data D , 1f(~ , w) is a prior distribution on the parameter space. The Bayesian predictive distribution p(xID) is obtained by averaging p(xl~, w) with respect to the posterior distribution p(~, wiD), and can be written as p(xID) = Jp(xl~, w)p(~, wID)d~dw. (22) The Bayes predictive distribution depends on the prior distribution 1f( ~, w) . As long as the prior is a smooth function, the first order asymptotic properties are the same for the mle and Bayes estimators in the regular case. However, at singularities, the situation is different. Here, we assume a uniform prior for w. For C we assume two different priors, the uniform prior and the Jeffreys prior. We show here a sketch of calculations in the case of Jeffreys prior, 7f(~,w) ex 1~ld . By introducing Id(u) = J ~ Iz + uldexp {_~Z2} dz, (23) after lengthy calculations, we obtain (24) where XT+! = ~(x + VTx) , Pd(x) = J Id(Y(w)) exp {~Y2(W)} dw. (25) Here Y(w) has the same form defined in (15), and Pd(x) is the function of the sufficient statistics x. By using the Edgeworth expansion, we have p(xID) (26) where \7 is the gradient and H2 (x) is the Hermite polynomial. We thus have the following theorem. Theorem 2. Under the Jeffreys prior for ~, the generalization error and the training error of the predictive distribution are given by Egen (27) Etr ain (28) Under the uniform prior, the above results hold by replacing Id(Y) in the definition of Pd(X) by 1. In addition, From (24), we can easily obtain Egen = (d + 1)/2T for the Jeffreys prior, and Egen = 1/2T for the uniform prior. The theorem shows rather surprising results: Under the uniform prior, the generalization error is constant and does not depend on d. This is completely different from the regular case. However, this striking result is given rise to by the uniform prior on f The uniform prior puts strong emphasis on the singularity, showing that one should be very careful for choosing a prior when the model includes singularities. In the case of J effreys prior, the generalization error increases in proportion to d, which is the same result as the regular case. In addition, the symmetric duality between Egen and E train does not hold for both of the uniform prior and the Jeffreys prior. 7 Gaussian random field of MLP In the case of MLP with one hidden unit , the log likelihood is written as 1 T 2 L(D;~, w)=-22:{Yt-~CPi9(w.Xt)} . t=l (29) Let us define a Gaussian random field depending on D and w, 1 Y(w) = 1m yT T LYt<P,6 (w? Xt) '"" N(O,A(w,w')) (30) t= l where A(w, w') = Ex [<p,6(w . x)<p,6 (w' . x)]. Theorem 3. For the mle, we have (31) Egen (32) Etrain (33) where A(w) = A(w, w). In order to analyze the Bayes predictive distribution , we define Sd(D,w) = 1 J A(w) d+1 Y (W)) { 1 y2 (w) } Id ( JA(W) exp --A() . A(w) 2 w (34) We then have the Edgeworth expansion of the predictive distribution of the form, p(Ylx, D) _1_ exp {_ y2 } {I ~ f(L y27f 2 + -'!L EW[V'Sd(D, w)<p,6 (w . x)] (35) Ew [Sd( D , w )] ~ EW[V'V'Sd(D,w)A(w)] H ( )} + 2T EW[Sd(D,w)] 2 Y , 1m yT where V' is the gradient with respect to Y(w). We thus have the following theorem. Theorem 4. Under the Jeffreys prior for ~ , the generalization error and the training error of the predictive distribution are given by Egen Etrain (36) = Under the uniform prior, the above results hold by redefining (37) We can also obtain Egen the uniform prior. = (d + 1)/2T for the Jeffreys prior, and Egen = 1/2T for There is a nice correspondence between the cone model and MLP. However, there is no sufficient statistics in the MLP case, while all the data are summarized in the sufficient statistics x in the cone model. 8 Conclusions and discussions We have analyzed the asymptotic behaviors of the MLE and Bayes estimators in terms of the generalization error and the training error by using simple statistical models (cone model and simple MLP), when the true parameter is at singularity. Since the classic paradigm of statistical inference based on the Cramer-Rao theorem does not hold in such a singular case, we need a new theory. The Gaussian random field has played a fundamental role. We can compare the estimation accuracy of the maximum likelihood estimator and the Bayesian predictive distribution from the results of analysis. Under the proposed framework , the various estimation methods can be studied and compared to each other. References [1] Saad, D. and Solla, S. A. (1995). Physical Review E, 52,4225-4243. [2] Amari, S. (1998). N eural Computation, 10,251-276. [3] Amari S. and Nagaoka, H. (2000). Methods of Information Geometry, AMS. [4] Amari, S., Park, H. , and Fukumizu, F. (2000). Neural Computation, 12, 13991409. [5] Park, H., Amari, S. and Fukumizu, F. (2000). Neural Networks, 13, 755-764. [6] Chen, A. M., Lu, H. , and Hecht-Nielsen, R. (1993). Neural Computations, 5, 910-927. [7] Riigger, S. M. and Ossen, A. (1997). Neural Processing Letters, 5, 63-72. [8] Fukumizu, K. and Amari, S. (2000) Neural Networks, 13 317-327. [9] Hagiwara, K. , Hayasaka, K. , Toda, N., Usui, S., and Kuno, K . (2001). Neural Networks, 14 1419-1430. [10] Watanabe, S. (2001). Neural Computation, 13, 899-933. [11] Fukumizu, K. (2001). Research Memorandum, 780, lnst. of Statistical Mathematics. [12] Dacunha-Castelle, D. and Gassiat, E. (1997). Probability and Statistics, 1,285317. [13] Hartigan, J. A. (1985). Proceedings of Berkeley Conference in Honor of J. Neyman and J. Kiefer, 2, 807-810.
2015 |@word effect:2 toda:1 normalized:1 version:1 true:4 polynomial:1 proportion:1 differ:1 y2:3 hence:4 classical:1 merged:1 symmetric:1 occurs:1 modifying:1 simulation:1 stochastic:1 eg:3 covariance:1 wid:3 gradient:3 subspace:1 shun:1 xid:4 ja:1 mapped:1 ld:1 reduction:1 criterion:2 fix:1 generalization:11 manifold:1 etr:1 singularity:15 wako:1 hold:8 z2:1 geometrical:2 surprising:1 cramer:3 si:2 exp:4 written:4 substituting:1 physical:2 shape:1 rise:2 jp:2 estimation:3 implying:1 observation:2 ain:1 wl:1 ozeki:1 neuromanifold:10 consistency:1 mathematics:1 iia:1 fukumizu:5 pointed:1 similarly:1 curved:1 situation:1 ron:1 gaussian:13 attack:1 apex:2 rather:1 hermite:1 sharp:1 c2:5 corollary:1 posterior:3 xtl:1 introduced:1 pair:1 specified:2 kl:1 redefining:1 likelihood:9 introduce:2 iwl:1 honor:1 behavior:7 am:1 subgroup:1 inference:2 brain:2 dynamical:1 vl:1 hidden:6 relation:8 determine:1 paradigm:5 summarize:1 becomes:2 lnst:1 including:2 maximizes:1 isotropy:1 critical:1 smooth:1 submanifolds:1 natural:2 etrain:7 calculation:2 usui:1 sphere:1 long:1 residual:1 field:8 equal:1 divided:1 having:2 mle:14 y:1 hecht:1 improve:1 berkeley:1 represents:1 park:3 lit:1 multilayer:7 expectation:1 demonstrates:1 normalization:1 whatever:2 unit:9 vh:1 prior:23 nice:1 review:1 addition:2 divergence:1 residue:1 asymptotic:4 embedded:1 treat:1 sd:6 limit:1 replaced:1 geometry:2 singular:7 permutation:1 saad:1 analyzing:1 id:4 mlp:9 subject:1 black:1 h2:1 emphasis:1 studied:1 sufficient:3 equivalence:3 mdl:2 principle:1 analyzed:1 easy:1 architecture:1 topology:1 edgeworth:2 bias:1 perceptron:2 computable:2 institute:1 taking:1 circle:1 empirical:1 dimension:3 regular:7 rao:3 algebraic:1 made:1 cannot:1 adaptive:1 selection:5 put:2 introducing:1 subset:1 clear:1 equivalent:4 saitama:1 uniform:10 yt:4 dz:1 ylx:3 go:2 reduced:1 exist:1 estimator:15 adaptively:1 density:1 fundamental:1 sign:1 estimated:3 search:1 dw:2 classic:1 write:1 coordinate:3 memorandum:1 iz:1 symmetry:1 elucidate:4 play:1 hirosawa:1 ichi:1 central:1 group:1 expansion:2 us:3 excellent:1 hartigan:2 vj:2 satisfying:1 noise:3 gassiat:1 asymptotically:3 toy:2 japan:1 account:1 cone:12 observed:1 role:2 wand:1 inverse:1 angle:1 summarized:1 letter:1 includes:1 striking:1 place:2 wj:2 reasonable:1 en:3 slow:1 vi:9 solla:1 depends:1 sub:1 watanabe:2 xl:3 analyze:2 pd:3 bayes:11 parallel:1 guaranteed:1 aic:1 spacetime:1 dynamic:1 quadratic:1 correspondence:1 played:1 trained:1 depend:2 nontrivial:1 theorem:9 occur:1 accuracy:1 predictive:14 kiefer:1 characteristic:2 percept:1 xt:5 showing:2 completely:1 hagiwara:2 eural:1 admits:1 po:5 easily:1 essential:1 bayesian:6 represented:2 various:1 alc:1 lu:1 riken:2 train:1 trajectory:1 hole:1 chen:1 plateau:3 neighborhood:1 choosing:1 synaptic:1 ed:3 lengthy:1 definition:1 wi:7 egen:11 amari:7 remarked:1 statistic:4 jeffreys:7 riemannian:1 nagaoka:1 invariant:2 ip:1 satisfies:2 wh:1 neyman:1 conditional:2 ubiquitous:2 identity:1 nielsen:1 careful:1 lyt:1 j2:1 end:1 fisher:5 change:1 gen:1 operation:1 degenerate:1 averaging:1 hierarchical:1 appropriate:1 called:1 strongly:1 pll:1 duality:1 kuno:1 regularity:1 sketch:1 vtx:1 perceptrons:5 replacing:1 ew:4 maximizer:2 depending:2 evaluate:2 ex:2 phenomenon:1 strong:1
1,114
2,016
ALGONQUIN - Learning dynamic noise models from noisy speech for robust speech recognition Brendan J. Freyl, Trausti T. Kristjansson l , Li Deng2 , Alex Acero 2 1 Probabilistic and Statistical Inference Group, University of Toronto http://www.psi.toronto.edu 2 Speech Technology Group , Microsoft Research Abstract A challenging, unsolved problem in the speech recognition community is recognizing speech signals that are corrupted by loud, highly nonstationary noise. One approach to noisy speech recognition is to automatically remove the noise from the cepstrum sequence before feeding it in to a clean speech recognizer. In previous work published in Eurospeech, we showed how a probability model trained on clean speech and a separate probability model trained on noise could be combined for the purpose of estimating the noisefree speech from the noisy speech. We showed how an iterative 2nd order vector Taylor series approximation could be used for probabilistic inference in this model. In many circumstances, it is not possible to obtain examples of noise without speech. Noise statistics may change significantly during an utterance, so that speechfree frames are not sufficient for estimating the noise model. In this paper, we show how the noise model can be learned even when the data contains speech. In particular, the noise model can be learned from the test utterance and then used to de noise the test utterance. The approximate inference technique is used as an approximate E step in a generalized EM algorithm that learns the parameters of the noise model from a test utterance. For both Wall Street J ournal data with added noise samples and the Aurora benchmark, we show that the new noise adaptive technique performs as well as or significantly better than the non-adaptive algorithm, without the need for a separate training set of noise examples. 1 Introduction Two main approaches to robust speech recognition include "recognizer domain approaches" (Varga and Moore 1990; Gales and Young 1996), where the acoustic recognition model is modified or retrained to recognize noisy, distorted speech, and "feature domain approaches" (Boll 1979; Deng et al. 2000; Attias et al. 2001; Frey et al. 2001), where the features of noisy, distorted speech are first denoised and then fed into a speech recognition system whose acoustic recognition model is trained on clean speech. One advantage of the feature domain approach over the recognizer domain approach is that the speech modeling part of the denoising model can have much lower com- plexity than the full acoustic recognition model. This can lead to a much faster overall system, since the denoising process uses probabilistic inference in a much smaller model. Also, since the complexity of the denoising model is much lower than the complexity of the recognizer, the denoising model can be adapted to new environments more easily, or a variety of denoising models can be stored and applied as needed. We model the log-spectra of clean speech, noise, and channel impulse response function using mixtures of Gaussians. (In contrast, Attias et al. (2001) model autoregressive coefficients.) The relationship between these log-spectra and the log-spectrum of the noisy speech is nonlinear, leading to a posterior distribution over the clean speech that is a mixture of non-Gaussian distributions. We show how a variational technique that makes use of an iterative 2nd order vector Taylor series approximation can be used to infer the clean speech and compute sufficient statistics for a generalized EM algorithm that can learn the noise model from noisy speech. Our method, called ALGONQUIN, improves on previous work using the vector Taylor series approximation (Moreno 1996) by modeling the variance of the noise and channel instead of using point estimates, by modeling the noise and channel as a mixture mixture model instead of a single component model, by iterating Laplace's method to track the clean speech instead of applying it once at the model centers, by accounting for the error in the nonlinear relationship between the log-spectra, and by learning the noise model from noisy speech. 2 ALGONQUIN's Probability Model For clarity, we present a version of ALGONQUIN that treats frames of log-spectra independently. The extension of the version presented here to HMM models of speech, noise and channel distortion is analogous to the extension of a mixture of Gaussians to an HMM with Gaussian outputs. Following (Moreno 1996), we derive an approximate relationship between the log spectra of the clean speech, noise, channel and noisy speech. Assuming additive noise and linear channel distortion, the windowed FFT Y(j) for a particular frame (25 ms duration, spaced at 10 ms intervals) of noisy speech is related to the FFTs of the channel H(j), clean speech 5(j) and additive noise N(j) by Y(j) = H(j)5(j) + N(j). (1) We use a mel-frequency scale, in which case this relationship is only approximate. However, it is quite accurate if the channel frequency response is roughly constant across each mel-frequency filter band. For brevity, we will assume H(j) = 1 in the remainder of this paper. Assuming there is no channel distortion simplifies the description of the algorithm. To see how channel distortion can be accounted for in a nonadaptive way, see (Frey et al. 2001). The technique described in this paper for adapting the noise model can be extended to adapting the channel model. Assuming H(j) = 1, the energy spectrum is obtained as follows: IY(j)1 2 = Y(j)*Y(j) = 5(j)* 5(j) + N(j)* N(j) + 2Re(N(j)* 5(j)) = 15(j)1 2 + IN(j)12 + 2Re(N(j)* 5(j)) , where "*,, denotes complex conjugate. If the phase of the noise and the speech are uncorrelated, the last term in the above expression is small and we can approximate the energy spectrum as follows: IYUW ~ ISUW + INUW? (2) Although we could model these spectra directly, they are constrained to be nonnegative. To make density modeling easier, we model the log-spectrum instead. An additional benefit to this approach is that channel distortion is an additive effect in the log-spectrum domain. Letting y be the vector containing the log-spectrum log IY(:W, and similarly for s and n , we can rewrite (2) as exp(y) ~ exp(s) + exp(n) = exp(s) 0 (1 + exp(n - s)) , where the expO function operates in an element-wise fashion on its vector argument and the "0" symbol indicates element-wise product. Taking the logarithm, we obtain a function gO that is an approximate mapping of sand n to y (see (Moreno 1996) for more details): y ~ g([~]) = s + In(l + exp(n - s)). (4) "T" indicates matrix transpose and InO and expO operate on the individual elements of their vector arguments. Assuming the errors in the above approximation are Gaussian, the observation likelihood is (5) p(yls,n) =N(y;g([~]),W), where W is the diagonal covariance matrix of the errors. A more precise approximation to the observation likelihood can be obtained by writing W as a function of s and n , but we assume W is constant for clarity. Using a prior p(s, n), the goal of de noising is to infer the log-spectrum of the clean speech s , given the log-spectrum ofthe noisy speech y. The minimum squared error estimate of sis s = Is sp(sly) , where p(sly) ex InP(yls, n)p(s, n). This inference is made difficult by the fact that the nonlinearity g([s n]T) in (5) makes the posterior non-Gaussian even if the prior is Gaussian. In the next section, we show how an iterative variational method that uses a 2nd order vector Taylor series approximation can be used for approximate inference and learning. We assume that a priori the speech and noise are independent - p(s , n) = p(s)p(n) - and we model each using a separate mixture of Gaussians. cS = 1, ... , NS is the class index for the clean speech and en = 1, ... ,Nn is the class index for the noise. The mixing proportions and Gaussian components are parameterized as follows: p(s) = LP(cS)p(slcS), p(C S) =7r~s , p(slc S) =N(s;JL~s ,~~s ), CS We assume the covariance matrices ~~s and ~~n are diagonal. Combining (5) and (6), the joint distribution over the noisy speech, clean speech class, clean speech vector, noise class and noise vector is p(y , s , cs, n , en) = N(y; g([~]), w)7r~sN(s; JL~s , ~~s )7r~N(n; JL~n , ~~n). (7) Under this joint distribution, the posterior p(s, nly) is a mixture of non-Gaussian distributions. In fact, for a given speech class and noise class, the posterior p(s, nics, en , y) may have multiple modes. So, exact computation of s is intractable and we use an approximation. 3 Approximating the Posterior For the current frame of noisy speech y, ALGONQUIN approximates the posterior using a simpler, parameterized distribution, q: p(s ,cS, n,cnly) ~ q(s,cS,n,c n ). The "variational parameters" of q are adjusted to make this approximation accurate, and then q is used as a surrogate for the true posterior when computing ? and learning the noise model (c.f. (Jordan et al. 1998)). For each cS and en, we approximate p(s, nics, en, y) by a Gaussian, (9) where 1J~'en and 1J~'en are the approximate posterior means of the speech and noise for classes cS and en, and <P ~~en, <P~.r;,n and <P~::'en specify the covariance matrix for the speech and noise for classes cS and en. Since rows of vectors in (4) do not interact and since the likelihood covariance matrix q, and the prior covariance matrices ~ ~. and ~~n are diagonal, the matrices <P~~ en, <P~.r;,n and <P~::'en are diagonal. The posterior mixing proportions for classes cS and en are q( cS , en) = Pc' en. The approximate posterior is given by q(s,n,cs,cn ) = q(s , nlcs ,cn)q(c S, en). The goal of variational inference is to minimize the relative entropy (KullbackLeibler divergence) between q and p: "''''11 ( K=~~ c' en s n ,cS,cn nI ) ). q s , n , cS ,cn) In q(s ( ,n S P s, c , n , c y This is a particularly good choice for a cost function, because, since lnp(y) doesn't depend on the variational parameters, minimizing K is equivalent to maximizing F = lnp () y - K "''''11 ( = ~~ e' en s n q s , n , cS ,cn) In p(s,cS,n,cn,y) ( S n) , q s, n, c ,c which is a lower bound on the log-probability of the data. So, variational inference can be used as a generalized E step (Neal and Hinton 1998) in an algorithm that alternatively maximizes a lower bound on lnp(y) with respect to the variational parameters and the noise model parameters, as described in the next section. Variational inference begins by optimizing the means and variances in (9) for each and en. Initially, we set the posterior means and variances to the prior means and variances. F does not have a simple form in these variational parameters. So, at each iteration, we make a 2nd order vector Taylor series approximation of the likelihood, centered at the current variational parameters, and maximize the resulting approximation to F. The updates are CS where g' 0 is a matrix of derivatives whose rows correspond to the noisy speech y and whose columns correspond to the clean speech and noise [s n]. The inverse posterior covariance matrix is the sum of the inverse prior covariance matrix and the inverse likelihood covariance matrix, modified by the Jacobian g' 0 for the mapping from s and n to y The posterior means are moved towards the prior means and toward values that match the observation y. These two effects are weighted by the inverse prior covariance matrix and the inverse likelihood covariance matrix. After iterating the above updates (in our experiments, 3 to 5 times) for each eS and en, the posterior mixing proportions that maximize :F are computed: where A is a normalizing constant that is computed so that L e.en Pe'en = 1. The minimum squared error estimate of the clean speech, s, is We apply this algorithm on a frame-by-frame basis, until all frames in the test utterance have been denoised. 4 Speed Since elements of s, nand y that are in different rows do not interact in (4), the above matrix algebra reduces to efficient scalar algebra. For 256 speech components, 4 noise components and 3 iterations of inference, our unoptimized C code takes 60 ms to denoise each frame. We are confident that this time can be reduced by an order of magnitude using standard implementation tricks. 5 Adapting the Noise Model Using Noisy Speech The version of ALGONQUIN described above requires that a mixture model of the noise be trained on noise samples, before the log-spectrum of the noisy speech can be denoised. Here, we describe how the iterative inference technique can be used as the E step in a generalized EM algorithm for learning the noise model from noisy speech. For a set of frames bound y(1), . .. , yeT) in a noisy test utterance, we construct a total :F = L:F(t) :::; Llnp(y(t)). t t The generalized EM algorithm alternates between updating one set of variational (t) ... T? h parameters Pe.en, 11 n(t) e'en, et c. ?or each f rame t=I, T ... , ,and maximizIng.r WIt respect to the noise model parameters 7r~n, J.t~n and ~~n. Since:F:::; Ltlnp(y(t)), this procedure maximizes a lower bound on the log-probability of the data. The use of the vector Taylor series approximations leads to an algorithm that maximizes an approximation to a lower bound on the log-probability of the data. Restaurant Street Airport Station Average dB 2.12 2.96 1.82 1.73 2.16 dB 3.87 4.78 2.27 3.24 3.54 dB 9.18 10.73 5.49 6.48 7.97 dB 20.51 13.52 14.97 15.18 18.54 o dB 47.04 45.68 36.00 37.24 41.49 -5dB 78.69 72.34 69.04 67.26 71.83 Average 16.54 17.53 12.11 12.77 14.74 Table 1: Word error rates (in percent) on set B of the Aurora test set, for the adaptive version of ALGONQUIN with 4 noise componentsset. 20 15 10 5 Setting the derivatives of :F with respect to the noise model parameters to zero, we obtain the following M step updates: ~nen +--- (~ ' " ' '~ " ' Pe. (t)en (opnn(t) e' e n t +d?lag (( 11e'n en (t) -#-t n)( (t) -#-ten n )T))) / ('"' en 11 n e' en ~ '"' ~ P (t) e. en ) cB t cB The variational parameters can be updated multiple times before updating the model parameters, or the variational parameters can updated only once before updating the model parameters. The latter approach may converge more quickly in some situations. 6 Experimental Results After training a 256-component speech model on clean speech, we used the adaptive version of ALGONQUIN to denoise noisy test utterances on two tasks: the publically available Aurora limited vocabulary speech recognition task (http://www.etsi.org/technicalactiv/dsr.htm); the Wall Street J ournal (WSJ) large vocabulary speech recognition task, with Microsoft's Whisper speech recognition system. We obtained results on all 48 test sets from partitions A and B of the Aurora database. Each set contains 24,000 sentences that have been corrupted from one of 4 different noise types and one of 6 different signal to noise ratios. Table 1 gives t he error rates for the adaptive version of ALGONQUIN, with 4 noise components. These error rates are superior to error rates obtained by our spectral subtraction technique for (Deng et al. 2000) , and highly competitive with other results on the Aurora task. Table 2 compares the performances of the adaptive version of ALGONQUIN and t he non-adaptive version. For the non-adaptive version, 20 non-speech frames are used to estimate the noise model. For the adaptive version, the parameters are init ialized using 20 non-speech frames and then 3 iterations of generalized EM are used to learn the noise model. The average error rate over all noise types and SNRs for set B of Aurora drops from 17.65% to 15.19% when t he noise adaptive algorithm is used to update the noise model. This is a relative gain of 13.94%. When 4 components are used there is a further gain of 2.5%. The Wall Street J ournal test set consists of 167 sentences spoken by female speakers. The Microsoft Whisper recognizer with a 5,000 word vocabulary was used to recognize these sentences. Table 2 shows that the adaptive version of algonquin . WER WER Reduction WER Reduction 20 frames 1 comp in WER 4 comps in WER Aurora, Set A 18.10% 15.91% 12.10% 15.62% 13.70% Aurora, Set B 14.74% 17.65% 15.19% 13.94% 16.49% WSJ, XD14, 10dB 30.00% 21.8% 27.33% 21.50% 28.33% WSJ, XD10, 10dB 21.80% 20.6% 20.6% 5.50'70 5.50 '70 Table 2: Word error rates (WER) and percentage reduction in WER for the Aurora test data and the Wall Street J ournal test data, without scaling. performs better than the non-adaptive version, especially on noise type "XD1 4" , which consists of the highly-nonstationary sound of a jet engine shutting down. For noise type "XD1O", which is stationary noise, we observe a gain, but we do not see any further gain for multiple noise components. 7 Conclusions A far as variational methods go, ALGONQUIN is a fast technique for denoising logspectrum or cepstrum speech feature vectors. ALGONQUIN improves on previous work using the vector Taylor series approximation, by using multiple component speech and noise models, and it uses an iterative variational method to produce accurate posterior distributions for speech and noise. By employing a generalized EM method, ALGONQUIN can estimate a noise model from noisy speech data. Our results show that the noise adaptive ALGONQUIN algorithm can obtain better results than the non-adaptive version. This is especially important for nonstationary noise, where the non-adaptive algorithm relies on an estimate of the noise based on a subset of the frames , but the adaptive algorithm uses all the frames of the utterance, even those that contain speech. A different approach to denoising speech features is to learn time-domain models. Attias et al. (2001) report results on a non-adaptive time-domain technique. Our results cannot be directly compared with theirs, since our results are for unscaled data. Eventually, the two approaches should be thoroughly compared. References Attias, H. , Platt , J . C., Acero, A., and Deng, L. 2001. Speech denoising and dereverberation using probabilistic models. In Advances in Neural Information Processing Systems 13. MIT Press, Cambridge MA. Boll, S. 1979. Suppression of acoustic noise in speech using spectral subtraction. IEEE Transactions on Acoustics, Speech and Signal Processing, 27:114- 120. Deng, L. , Acero, A., Plumpe, M., and Huang, X. D. 2000. Large-vocabulary speech recognition under adverse acoustic environments. In Proceedings of the International Conference on Spoken Language Processing, pages 806- 809. Frey, B. J. , Deng, L. , Acero, A., and Kristjansson, T. 2001. ALGONQUIN: Iterating Laplace's method to remove multiple types of acoustic distortion for robust speech recognition. In Proceedings of Eurospeech 2001. Gales, M. J. F. and Young, S. J . 1996. Robust continuous speech recognition using parallel model combination. IEEE Speech and Audio Processing, 4(5):352- 359. Jordan, M. 1. , Ghahramani, Z., J aakkola, T. S., and Saul , L. K. 1998. An introduction to variational methods for graphical models. In Jordan, M. 1., editor, Learning in Graphical Models. Kluwer Academic Publishers, Norwell MA. Moreno, P. 1996. Speech R ecognition in Noisy Environments. Carnegie Mellon University, Pittsburgh PA. Doctoral dissertation. Neal, R. M. and Hinton, G. E. 1998. A view of the EM algorithm that justifies incremental, sparse, and other variants . In Jordan , M. 1. , editor, Learning in Graphical Models , pages 355- 368. Kluwer Academic Publishers, Norwell MA . Varga, A. P. and Moore, R. K. 1990. Hidden Markov model decomposition of speech and noise. In Proceedings of th e International Conference on Acoustics, Speech and Signal Processing, pages 845- 848. IEEE Press.
2016 |@word version:13 proportion:3 nd:4 kristjansson:2 accounting:1 covariance:10 decomposition:1 reduction:3 series:7 contains:2 current:2 com:1 si:1 yet:1 additive:3 partition:1 remove:2 moreno:4 drop:1 update:4 stationary:1 dissertation:1 toronto:2 org:1 simpler:1 windowed:1 consists:2 roughly:1 automatically:1 begin:1 estimating:2 maximizes:3 spoken:2 platt:1 before:4 frey:3 treat:1 whisper:2 doctoral:1 challenging:1 limited:1 procedure:1 significantly:2 adapting:3 word:3 inp:1 cannot:1 noising:1 acero:4 applying:1 writing:1 www:2 equivalent:1 center:1 maximizing:2 go:2 independently:1 duration:1 wit:1 laplace:2 analogous:1 updated:2 exact:1 us:4 trick:1 element:4 pa:1 recognition:14 particularly:1 updating:3 database:1 trausti:1 environment:3 complexity:2 dynamic:1 trained:4 depend:1 rewrite:1 algebra:2 basis:1 easily:1 joint:2 htm:1 snrs:1 fast:1 describe:1 whose:3 quite:1 lag:1 distortion:6 statistic:2 noisy:21 sequence:1 advantage:1 product:1 remainder:1 combining:1 mixing:3 description:1 moved:1 produce:1 wsj:3 incremental:1 derive:1 c:17 filter:1 centered:1 sand:1 feeding:1 wall:4 adjusted:1 extension:2 exp:6 cb:2 mapping:2 purpose:1 recognizer:5 weighted:1 mit:1 gaussian:8 modified:2 indicates:2 likelihood:6 contrast:1 brendan:1 suppression:1 inference:11 publically:1 nn:1 aurora:9 initially:1 nand:1 hidden:1 unoptimized:1 overall:1 priori:1 constrained:1 airport:1 once:2 noisefree:1 construct:1 report:1 recognize:2 divergence:1 individual:1 phase:1 microsoft:3 highly:3 mixture:8 pc:1 algonquin:16 accurate:3 norwell:2 taylor:7 logarithm:1 re:2 column:1 modeling:4 cost:1 subset:1 recognizing:1 eurospeech:2 kullbackleibler:1 stored:1 corrupted:2 combined:1 confident:1 thoroughly:1 density:1 international:2 probabilistic:4 iy:2 quickly:1 squared:2 containing:1 huang:1 gale:2 derivative:2 leading:1 li:1 de:2 slc:1 coefficient:1 view:1 competitive:1 denoised:3 parallel:1 minimize:1 ni:1 variance:4 spaced:1 ofthe:1 correspond:2 comp:2 published:1 energy:2 frequency:3 psi:1 unsolved:1 gain:4 improves:2 response:2 specify:1 cepstrum:2 sly:2 until:1 nonlinear:2 mode:1 impulse:1 effect:2 contain:1 true:1 moore:2 neal:2 during:1 speaker:1 mel:2 m:3 generalized:7 plexity:1 performs:2 percent:1 variational:16 wise:2 superior:1 jl:3 he:3 approximates:1 kluwer:2 theirs:1 mellon:1 cambridge:1 similarly:1 nonlinearity:1 language:1 posterior:15 showed:2 female:1 optimizing:1 yls:2 nics:2 lnp:3 minimum:2 additional:1 deng:5 subtraction:2 converge:1 maximize:2 signal:4 full:1 multiple:5 sound:1 infer:2 reduces:1 faster:1 academic:2 match:1 jet:1 variant:1 circumstance:1 iteration:3 interval:1 publisher:2 operate:1 db:8 jordan:4 nonstationary:3 fft:1 variety:1 restaurant:1 simplifies:1 cn:6 attias:4 expression:1 speech:72 varga:2 iterating:3 band:1 ten:1 reduced:1 http:2 percentage:1 track:1 carnegie:1 group:2 clarity:2 clean:16 nonadaptive:1 sum:1 inverse:5 parameterized:2 wer:7 distorted:2 scaling:1 bound:5 nonnegative:1 adapted:1 alex:1 expo:2 speed:1 argument:2 freyl:1 alternate:1 combination:1 conjugate:1 smaller:1 across:1 em:7 lp:1 eventually:1 needed:1 letting:1 fed:1 available:1 gaussians:3 nen:1 apply:1 observe:1 spectral:2 denotes:1 include:1 graphical:3 ghahramani:1 especially:2 approximating:1 added:1 diagonal:4 surrogate:1 ecognition:1 separate:3 street:5 hmm:2 toward:1 assuming:4 code:1 index:2 relationship:4 ratio:1 minimizing:1 difficult:1 implementation:1 observation:3 markov:1 benchmark:1 situation:1 extended:1 ino:1 precise:1 hinton:2 frame:14 station:1 retrained:1 community:1 rame:1 boll:2 sentence:3 acoustic:8 engine:1 learned:2 dereverberation:1 dsr:1 nly:1 xd1:1 technology:1 utterance:8 sn:1 prior:7 relative:2 sufficient:2 editor:2 uncorrelated:1 unscaled:1 row:3 accounted:1 last:1 transpose:1 saul:1 taking:1 sparse:1 benefit:1 vocabulary:4 autoregressive:1 doesn:1 made:1 adaptive:17 far:1 employing:1 transaction:1 approximate:10 pittsburgh:1 alternatively:1 spectrum:15 continuous:1 iterative:5 table:5 channel:12 learn:3 robust:4 init:1 interact:2 complex:1 shutting:1 domain:7 sp:1 main:1 noise:63 denoise:2 en:30 fashion:1 n:1 pe:3 jacobian:1 ffts:1 learns:1 young:2 down:1 symbol:1 normalizing:1 intractable:1 magnitude:1 justifies:1 easier:1 entropy:1 scalar:1 loud:1 relies:1 ma:3 goal:2 towards:1 change:1 adverse:1 operates:1 denoising:8 called:1 total:1 e:1 experimental:1 latter:1 brevity:1 audio:1 ex:1
1,115
2,017
Stabilizing Value Function with the Xin Wang Department of Computer Science Oregon State University Corvallis, OR, 97331 wangxi@cs. orst. edu Thomas G Dietterich Department of Computer Science Oregon State University Corvallis, OR, 97331 tgd@cs. orst. edu Abstract We address the problem of non-convergence of online reinforcement learning algorithms (e.g., Q learning and SARSA(A)) by adopting an incremental-batch approach that separates the exploration process from the function fitting process. Our BFBP (Batch Fit to Best Paths) algorithm alternates between an exploration phase (during which trajectories are generated to try to find fragments of the optimal policy) and a function fitting phase (during which a function approximator is fit to the best known paths from start states to terminal states). An advantage of this approach is that batch value-function fitting is a global process, which allows it to address the tradeoffs in function approximation that cannot be handled by local, online algorithms. This approach was pioneered by Boyan and Moore with their GROWSUPPORT and ROUT algorithms. We show how to improve upon their work by applying a better exploration process and by enriching the function fitting procedure to incorporate Bellman error and advantage error measures into the objective function. The results show improved performance on several benchmark problems. 1 Introduction Function approximation is essential for applying value-function-based reinforcement learning (RL) algorithms to solve large Markov decision problems (MDPs). However, online RL algorithms such as SARSA(A) have been shown experimentally to have difficulty converging when applied with function approximators. Theoretical analysis has not been able to prove convergence, even in the case-of linear function approximators. (See Gordon (2001), however, for a non-divergence result.) The heart of the problem is that the approximate values of different states (e.g., 81 and 82) are coupled through the parameters of the function approximator. The optimal policy at state 81 may require increasing a parameter, while the optimal policy at state 82 may require decreasing it. As a result, algorithms based on local parameter updates tend to oscillate or even to diverge. To avoid this problem, a more global approach is called for-an approach that can consider Sl and S2 simultaneously and find a solution that works well in both states. One approach is to formulate the reinforcement learning problem as a global search through a space of parameterized policies as in the policy gradient algorithms (Williams, 1992; Sutton, McAllester, Singh, & Mansour, 2000; Konda & Tsitsiklis, 2000; Baxter & Bartlett, 2000). This avoids the oscillation problem, but the resulting algorithms are slow and only converge to local optima. We pursue an alternative approach that formulates the function approximation problem as a global supervised learning problem. This approach, pioneered by Boyan and Moore in their GROWSUPPORT (1995) and ROUT (1996) algorithms, separates the reinforcement learning problem into two subproblems: the exploration problem (finding a good partial value function) and the representation problem (representing and generalizing that value function). These algorithms alternate between two phases. During the exploration phase, a support set of points is constructed whose optimal values are known within some tolerance. In the function fitting phase, a function approximator V is fit to the support set. In this paper, we describe two ways of improving upon GROWSUPPORT and ROUT. First, we replace the support set with the set of states that lie along the best paths found during exploration. Second, we employ a combined error function that includes terms for the supervised error, the Bellman error, and the advantage error (Baird, 1995) into the function fitting process. The resulting BFBP (Batch Fit to Best Paths) method gives significantly better performance on resource-constrained scheduling problems as well as on the mountain car toy benchmark problem. 2 GrowSupport, ROUT, and BFBP Consider a deterministic, episodic MDP. Let s' == a(s) denote the state s' that results from performing a in s and r(a, s) denote the one-step reward. Both GROWSUPPORT and ROUT build a support set S == {(Si' V(Si))} of states whose optimal values V (s) are known with reasonable accuracy. Both algorithms initialize S with a set of terminal states (with V(s) == 0). In each iteration, a function approximator V is fit to S to minimize :Ei[V(Si) - V(Si)]2. Then, an exploration process attempts to identify new points to include in S. In GROWSUPPORT, a sample of points X is initially drawn from the state space. In each iteration, after fitting V, GROWSUPPORT computes a new estimate V(s) for each state sEX according to V(s) == max a r(s, a) + V(a(s)), where V(a(s)) is computed by executing the greedy policy with respect to V starting in a(s). If V(a(s)) is within c of V(a(s)), for all actions a, then (s, V(s)) is added to S. ROUT employs a different procedure suitable for stochastic MDPs. Let P(s'ls, a) be the probability that action a in state s results in state s' and R(s'ls, a) be the expected one-step reward. During the exploration phase, ROUT generates a trajectory from the start state to a terminal state and then searches for a state s along that trajectory such that (i) V(s) is not a good approximation to the backedup value V(s) == maxa :Est P(s'ls, a)[R(s'ls, a) + V(s')], and (ii) for every state s along a set of rollout trajectories starting at s', V(s) is within c of the backed-up value maxa :Est P(s'ls, a)[R(s'ls, a) + V(s')]. If such a state is found, then (s, V(s)) is added to S. Both GROWSUPPORT and ROUT rely on the function approximator to generalize well at the boundaries of the support set. A new state s can only be added to S if V has generalized to all of s's successor states. H this occurs consistently, then eventually the support set will expand to include all of the starting states of the MDP, at which point a satisfactory policy has been found. However, if this "boundary generalization" does not occur, then no new points will be added to S, and both GROWSUPPORT and ROUT. terminate without a solution. Unfortunately, most regression methods have high bias and variance near the boundaries of their training data, so failures of boundary generalization are common. These observations led us to develop the BFBP algorithm. In BFBP, the exploration process maintains a data structure S that stores the best known path from the start state to a terminal state and a "tree" of one-step departures from this best path (Le., states that can be reached by executing an action in some state on the best path). At each state Si E S, the data structure stores the action executed in that state (to reach the next state in the path), the one-step reward ri, and the estimated value V(Si). S also stores each action a_ that causes a departure from the best path along with the resulting state S_, reward r_ and estimated value V(s_). We will denote by B the subset of S that constitutes the best path. The estimated values V are computed as folloV1S. For states S'i E B, V(Si) is computed 'by summing the immediate rewards r j for all steps j 2: i along B. For the one-step departure states s_, V(s_) is computed from an exploration trial in which the greedy policy was followed starting in state s_. at fuitially, S is empty, so a random trajectory is generated from the start state So to a terminal state, and it becomes the initial best known path. fu subsequent iterations, is chosen and executed to a state Si E B is chosen at random, and an action 1= produce state s' and reward r'. Then the greedy policy (with respect to the current V) is executed until a terminal state is reached. The rewards along this new path are summed to produce V(s'). If V(s') +r' > V(Si), then the best path is revised as follows. The new best action in state Si becomes al with estimated value V(s') +r' . This improved value is then propagated backwards to update the V estimates for in state Si becomes an inferior all ancestor states in B. The old best action action a_ with result state s_. Finally all descendants of s_ along the old best path are deleted. This method of investigating one-step departures from the best path is inspired by Harvey and Ginsberg's (1995) limited discrepancy search (LDS) algorithm. In each exploration phase, K one-step departure paths are explored. a' at at After the exploration phase, the value function approximation V is recomputed with the goal of minimizing a combined error function: J(V) == As L (V(s) - V(S))2 + Ab L (V(s) sES Aa L L [r(s, a*) + V(a*(s))])2 + sEB ([r(s,a-) + V(a-(s))] - [r(s,a*) + V(a*(s))]):. The three terms of this objective function are referred to as the supervised, Bellman, and advantage terms. Their relative importance is controlled by the coefficients As, Ab' and Au. The supervised term is the usual squared error between the V(s) values stored in S and the fitted values V(s). The Bellman term is the squared error between the fitted value and the backed-up value of the next state on the best path. And the advantage term penalizes any case where the backed-up value of an inferior action a_ is larger than the backed-up value of the best action a*. The notation (u)+ == u if u 2: 0 and 0 otherwise. TheoreIll 1 Let M be a deterministic MDP such that (aJ there are only a finite number of starting states, (bJ there are only? a finite set of actions executable in each state, and (c) all policies reach a terminal state. Then BFBP applied to M converges. Proof: The LDS exploration process is monotonic, since the data structure S is only updated if a new best path is found. The conditions of the theorem imply that there are only a finite number of possible paths that? can be explored from the starting states to the terminal states. Hence, the data structure S will eventually converge. Consequently, the value function V fit to S will also converge. Q.E.D. The theorem requires that the MDP contain no cycles. There are cycles in our jobshop scheduling problems, but we eliminate them by remembering all states visited along the current trajectory and barring any action that would return to a previously visited state. Note also that the theorem applies to MDPs with continuous state spaces provided the action space and the start states are finite. Unfortunately, BFBP does not necessarily converge to an optimal policy. This is because LDS exploration can get stuck in a local optimum such that all one step departures using the V-greedy policy produce trajectories that do not improve over the current best path. Hence, although BFBP resembles policy iteration, it does not have the same optimality guarantees,. because policy iteration evaluates the current greedy policy in all states in the state space. Theoretically, we could prove convergence to the optimal policy under modified conditions. If we replace LDS exploration with ?-greedy exploration, then exploration will converge to the optimal paths with probability 1. When trained on those paths, if the function approximator fits a sufficiently accurate V, then BFBS will converge optimally. hI our experiments, however, we have found that ?-greedy gives no improvement over LDS, whereas LDS exploration provides more complete coverage of one-step departures from the current best path, and these are used in J(V). 3 Experimental Evaluation We have studied five domains: Grid World and Puddle World (Boyan & Moore, 1995), Mountain Car (Sutton, 1996), and resource-constrained scheduling problems ART-1 and ART-2 (Zhang & Dietterich, 1995). For the first three domains, following Boyan and Moore, we compare BFBP with GROWSUPPORT. For the final domain, it is difficult to draw a sample of states X from the state space to initialize GROWSUPPORT. Hence, we compare against ROUT instead. As mentioned above, we detected and removed cycles from the scheduling domain (since ROUT requires this). We retained the cycles in the first three problems. On mountain car, we also applied SARSA(A) with the CMAC function approximator developed by Sutton (1996). We experimented with two function approximators: regression trees (RT) and locally-weighted linear regression (LWLR). Our regression trees employ linear separating planes at the internal nodes and linear surfaces at the leaf nodes. The trees are grown top-down in the usual fashion. To determine the splitting plane at a node, we choose a state Si at random from S, choose one of its inferior children S_, and construct the plane that is the perpendicular bisector of these two points. The splitting plane is evaluated by fitting the resulting child nodes to the data (as leaf nodes) and computing the value of J (V). A number C of parent-child pairs (Si' S - ) are generated and evaluated, and the best one is retained to be the splitting plane. This process is then repeated recursively until a node contains fewer than M data points~ The linear surfaces at the leaves are trained by gradient descent to minimize J(V). The gradient descent terminates after 100 steps or earlier if J becomes very small. In our experiments, we tried all combinations of the following parameters and report the best results: (a) 11 learning rates (from 0.00001 to 0.1), (b) M == 1, Table 1: Comparison of results on three toy domains. Problem Domain Grid World Algorithms GROWSUPPORT BFBP Puddle World G ROWSUPPORT Mountain Car BFBP SARSA(A) GROWSUPPORT BFBP Optimal Policyfj Yes Yes Yes Yes No No Yes Best Policy Length 39 39 39 39 103 93 88 Table 2: Results of ROUT and BFBP on scheduling problem ART-I-TRNOO I Performance I ROUT (RT) I ROUT (LWLR) I BFBP (RT) I Best policy explored I 1.75 I 1.55 I 1.50 I Best final learned policy I 1.8625 I 1.8125 I 1.55 10, 20, 50, 100, 1000, (c) C == 5, 10, 20, 50, 100, and (d) K == 50, 100, 150, 200. For locally-weighted linear regression, we replicated the methods of B'oyan and Moore. To compute V(s), a linear regression is performed using all points Si E S weighted by their distance to S according to the kernel exp -(Ilsi - sII 2 /a 2 ). We experimented with all combinations of the following parameters and report the best results: (a) 29 values (from 0.01 to 1000.0) of the tolerance E that controls the addition of new points to S, and (b) 39 values (from 0.01 to 1000.0) for a. We execute ROUT and GROWSUPPORT to termination. We execute BFBP for 100 iterations, but it converges much earlier: 36 iterations for the grid world, 3 for puddle world, 10 for mountain car, and 5 for the job-shop scheduling problems. Table 1 compares the results of the algorithms on the toy domains with parameters for each method tuned to give the best results and with As == 1 and Ab == Aa == o. In all cases, BFBP matches or beats the other methods. In Mountain Car, in particular, we were pleased that BFBP discovered the optimal policy very quickly. Table 2 compares the results of ROUT and BFBP on job-shop scheduling problem TRNOO from problem set ART-1 (again with As == 1 and Ab == Aa == 0). For ROUT, results with both LWLR and RT are shown. LWLR gives better results for ROUT. We conjecture that this is because ROUT needs a value function approximator that is conservative near the boundary of the training data, whereas BFBP does not. We report both the best policy found during the iterations and the final policy at convergence. Figure 1 plots the r,esults for ROUT (LWLR) against BFBP (RT) for eight additional scheduling problems from ART-I. The figure of merit is RDF, which is a normalized measure of schedule length (small values are preferred). BFBP's learned policy out-performs ROUT's in every case. The experiments above all employed only the supervised term in the error function J. These experiments demonstrate that LDS exploration gives better training sets than the support set methods of GROWSUPPORT and ROUT. Now we turn to the question of whether the Bellman and advantage terms can provide improved results. For the grid world and puddle world tasks, the supervised term already gives optimal performance, so we focus on the mountain car and job-shop scheduling problems. Table 3 summarizes the results for BFBP on the mountain car problem. All parameter settings, except for the last, succeed in finding the optimal policy. To get 2.4 best policy explored + y=xbest finalleamed policy x 2.2 G:' Q es <l) ? Xx 1.8 + x x ? ~ 0.- 1.6 + 0... ffP=l + 1.4 1.2 1 1 1.4 1.2 1.6 1.8 2.2 2.4 ROUT performance (RDF) Figure 1: Performance of Rout vs. BFBP over 8 job shop scheduling problems Table 3: Fraction of parameter settings that give optimal performance for BFBP on the mountain car problem .As 0.0 1.0 1.0 1.0 .Ab 0.0 0.0 10.0 100.0 .Aa 1.0 0.0 0.0 0.0 # settings 2/1311 52/1280 163/1295 4/939 As 0.0 1.0 1.0 1.0 Ab 1.0 0.0 0.0 1000.0 Aa 0.0 10.0 100.0 0.0 # settings 1/1297 184/1291 133/1286 0/1299 a sense of the robustness of the method, we also report the fraction of parameter settings that gave the optimal policy. The number of parameter settings tested (the denominator) should be the same for all combinations of A values. Nonetheless, for reasons unrelated to the parameter settings, some combinations failed to be executed by our distributed process scheduler. The best settings combine As == 1 with either Ab == 10 or Aa == 10. However, if we employ either the Bellman or the advantage term alone, the results are poor. Hence, it appears that the supervised term is very important for good performance, but that the advantage and Bellman terms can improve performance substantially .and reduce the sensitivity of BFBP to the settings of the other parameters. Table 4 shows the performance of BFBP on ART-I-TRNOO. The best performance (at convergence) is obtained with As == Aa == 1 and Ab == O. As with mountain car, these experiments show that the supervised term is the most important, but that it gives even better results when combined with the advantage term. All of the above experiments compare performance on single problems. We also tested the ability of BFBP to generalize to similar problems following the formulation of (Zhang & Dietterich, 1995). Figure 2 compares the performance of neural networks and regression trees as function approximators for BFBP. Both were trained on job shop scheduling problem set ART-2. Twenty of the problems in ART-2 were used for training, 20 for cross-validation, and 50 for testing. Eleven different values for As, Ab' Aa and eight different values for the learning rate were tried, with the best values selected according to the cross-validation set. Figure 2 shows that BFBP is significantly better than the baseline performance (with RDF Table 4: Performance ofBFBP on ART-1-TENOO for different settings of the .A parameters. The ('perform;' column gives the best RDF in any iteratIon and the RDF at convergence. .A 8 0 0 0 1 1 .Ab .Aa 0 1 10 0 1 1 1 1 1 1 .A 8 0 0 0 1 1 perform . 1.50/1.75 1.50/1.775 1.50/1.775 1.50/1.488 1.525/1.55 .Ab .Aa 1 1 100 0 1 0 10 1 10 10 perform. 1.50/1.775 1.50/1.825 1.50/1.738 1.463/1.525 1.50/1.588 .A 8 1 0 1 1 1 .Ab .Aa 0 1 1 0 1 0 100 a 100 100 perform. 1.50/1.55 1.50/1.65 1.50/1.563 1.50/1.588 1.50/1.675 1.8 . - - - - - - , - - - - . - - - - - - - , . - - - . . . , - - - - , - - - - - - , BFBP neural net ----*---. BFBP regression tree --- -G---- 1.75 _________________________RP_E _ 1.7 1.65 LL o 1.6 0: ~ ~ 1.55 ~ 1.45 1.4'?._?4<.,_. ,'"'::-..... ,c ._?? _._._.-!n-._,._. . ...._._.,.._._._. ._,._. . ._., ._.._._. ,,_._. ._.,_T._..O_L__ . . 1.35 1.3 L . . - -_ _- - - l - -_ _- - ' - -_ _- - L o 10 15 - ' - - -_ _- - - ' -_ _-----' 20 25 30 LOS iteration Figure 2: BFBP on ART-2 using neural nets and regression trees. "RDF" is a hand-coded heuristic, "TDL" is Zhang's TD(.A) neural network. as a search heuristic) and that its performance is comparable to TD(A) with neural networks (Zhang & Dietterich, 1995). Figure 3 shows that for ART-2, using parent/inferior-child pair splits gives better results than using axis-parallel splits. 4 Conclusions This paper has shown that the exploration strategies underlying GROWSUPPORT and ROUT can be improved by simply remembering and training on the best paths found between start and terminal states. Furthermore, the paper proved that the BFBP method converges for arbitrary function approximators, which is a result that has not yet been demonstrated for online methods such as SARSA(A). In addition, we have shown that the performance of our BFBP algorithm can be further improved (and made more robust) by incorporating a penalty for violations of the Bellman equation or a penalty for preferring inferior actions (an advantage error). Taken together, these results show that incremental-batch value function approximation can be a reliable, convergent method for solving deterministic reinforcement learning problems. The key to the success of the method is the ability to separate the exploration process from the function approximation process and to make the exploration process convergent. This insight should also be applicable to stochastic episodic MDPs. 1.9 , - - - - - - - . - - - - - , - - - - - - , - - - - . . , - - - - - - - , - - - - - , *... axis-parallel .... parent/inferior-child .. ??11 ???. 1.3 L -_ _- - l - o - ' - - -_ _- - - ' - - L -_ _---L._ _- - - - I 10 20 15 25 30 LDS iteration Figure 3: Axis parallel splits versus parent/inferior-child pair splits on ART-2 Acknowledgments The authors gratefully acknowledge the support of AFOSR under contract F4962098-1-0375, and the NSF under grants IRl-9626584, I1S-0083292, 1TR-5710001197, and EIA-9818414. We thank Valentina Zubek for her careful reading of the paper. References Baird, L. C. (1995). Residual algorithms: Reinforcement learning with function approximation. In ICML-95, 30-37, San Francisco, CA. Morgan Kaufmann. Baxter, J., & Bartlett, P. L. (2000). Reinforcement learning in POMDP's via direct gradient ascent. In ICML-2000, 41-48. Morgan Kaufmann, San Francisco, CA. Boyan, J. A., & Moore, A. W. (1995). Generalization in reinforcement learning: Safely approximating the value function. In NIPS-7, 369-376. The MIT Press, Cambridge. Boyan, J. A., & Moore, A. W. (1996). Learning evaluation functions for large acyclic domains. In ICML-96, 63-70. Morgan Kaufmann, San Francisco, CA. Gordon, G. J. (2001). Reinforcement learning with function approximation converge to a region. In NIPS-13, 1040-1046. The MIT Press. Harvey, W. D., & Ginsberg, L. P. (1995). Limited discrepancy search. In IJCAI-95, 825-830. Morgan Kaufmann. Konda, V. R., & Tsitsiklis, J. N. (2000). Policy gradient methods for reinforcement learning with function approximation. In NIPS-12, 1008-1014 Cambridge, MA. MIT Press. Moll, R., Barto, A. G., Perkins, T. J., & Sutton, R. S. (1999). Learning instanceindependent value functions to enhance local search. In NIPS-ll, 1017-1023. Sutton, R. S., McAllester, D., Singh, S., & Mansour, Y. (2000). Policy gradient methods for reinforcement learning with function approximation. In NIPS-12, 1057-1063. Sutton, R. S. (1996). Generalization in reinforcement learning: Successful examples using sparse coarse coding. In NIPS-8, 1038-1044. The MIT Press, Cambridge. Williams, R. J. (1992). Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning, 8,229. -.. Zhang, W., & Dietterich, T. G. (1995). A reinforcement learning approach to job-shop scheduling. In IJCAI-95, 1114-1120. Morgan Kaufmann, San Francisco, CA.
2017 |@word trial:1 sex:1 termination:1 tried:2 tr:1 recursively:1 initial:1 contains:1 fragment:1 tuned:1 current:5 ginsberg:2 si:14 yet:1 subsequent:1 eleven:1 plot:1 update:2 v:1 alone:1 greedy:7 leaf:3 fewer:1 selected:1 plane:5 provides:1 coarse:1 node:6 zhang:5 five:1 rollout:1 along:8 constructed:1 sii:1 direct:1 descendant:1 prove:2 tdl:1 fitting:8 combine:1 theoretically:1 expected:1 terminal:9 bellman:8 inspired:1 decreasing:1 td:2 increasing:1 becomes:4 provided:1 xx:1 notation:1 unrelated:1 underlying:1 zubek:1 mountain:10 pursue:1 substantially:1 maxa:2 developed:1 finding:2 guarantee:1 safely:1 every:2 control:1 grant:1 local:5 sutton:6 path:24 au:1 resembles:1 studied:1 xbest:1 limited:2 enriching:1 perpendicular:1 acknowledgment:1 testing:1 procedure:2 cmac:1 episodic:2 significantly:2 get:2 cannot:1 seb:1 scheduling:12 applying:2 deterministic:3 demonstrated:1 backed:4 williams:2 starting:6 l:6 pomdp:1 formulate:1 stabilizing:1 splitting:3 insight:1 updated:1 pioneered:2 lwlr:5 wang:1 region:1 cycle:4 removed:1 mentioned:1 reward:7 trained:3 singh:2 solving:1 upon:2 grown:1 describe:1 detected:1 whose:2 heuristic:2 larger:1 solve:1 s:1 otherwise:1 ability:2 final:3 online:4 advantage:10 net:2 los:1 convergence:6 empty:1 optimum:2 parent:4 ijcai:2 produce:3 incremental:2 executing:2 converges:3 develop:1 job:6 coverage:1 c:2 stochastic:2 exploration:22 mcallester:2 successor:1 require:2 generalization:4 sarsa:5 sufficiently:1 exp:1 bj:1 applicable:1 visited:2 weighted:3 mit:4 modified:1 r_:1 avoid:1 barto:1 focus:1 improvement:1 consistently:1 baseline:1 sense:1 eliminate:1 initially:1 her:1 ancestor:1 expand:1 constrained:2 summed:1 initialize:2 art:12 construct:1 barring:1 icml:3 constitutes:1 discrepancy:2 report:4 connectionist:1 gordon:2 employ:4 simultaneously:1 divergence:1 phase:8 attempt:1 ab:12 evaluation:2 violation:1 tgd:1 accurate:1 fu:1 partial:1 tree:7 old:2 penalizes:1 theoretical:1 fitted:2 column:1 earlier:2 formulates:1 subset:1 i1s:1 successful:1 optimally:1 stored:1 combined:3 sensitivity:1 preferring:1 contract:1 diverge:1 enhance:1 together:1 quickly:1 squared:2 again:1 choose:2 return:1 toy:3 coding:1 includes:1 coefficient:1 baird:2 oregon:2 performed:1 try:1 reached:2 start:6 maintains:1 parallel:3 minimize:2 accuracy:1 variance:1 kaufmann:5 identify:1 yes:5 generalize:2 lds:8 trajectory:7 reach:2 failure:1 evaluates:1 against:2 nonetheless:1 proof:1 propagated:1 proved:1 car:10 schedule:1 appears:1 supervised:8 improved:5 formulation:1 evaluated:2 execute:2 eia:1 furthermore:1 until:2 hand:1 ei:1 irl:1 aj:1 mdp:4 dietterich:5 contain:1 normalized:1 hence:4 moore:7 satisfactory:1 ffp:1 ll:2 during:6 inferior:7 generalized:1 complete:1 demonstrate:1 performs:1 theoreill:1 common:1 executable:1 rl:2 corvallis:2 cambridge:3 grid:4 gratefully:1 surface:2 store:3 harvey:2 success:1 approximators:5 morgan:5 additional:1 remembering:2 employed:1 converge:7 determine:1 ii:1 match:1 cross:2 coded:1 controlled:1 converging:1 regression:9 denominator:1 iteration:11 kernel:1 adopting:1 whereas:2 addition:2 ascent:1 tend:1 near:2 backwards:1 split:4 baxter:2 fit:7 gave:1 moll:1 jobshop:1 reduce:1 valentina:1 tradeoff:1 whether:1 handled:1 bartlett:2 penalty:2 oscillate:1 cause:1 action:15 locally:2 sl:1 nsf:1 estimated:4 recomputed:1 key:1 drawn:1 deleted:1 fraction:2 rout:25 parameterized:1 reasonable:1 oscillation:1 draw:1 decision:1 summarizes:1 comparable:1 hi:1 followed:1 convergent:2 occur:1 perkins:1 ri:1 generates:1 optimality:1 performing:1 conjecture:1 department:2 according:3 alternate:2 combination:4 poor:1 terminates:1 heart:1 taken:1 resource:2 equation:1 previously:1 turn:1 eventually:2 merit:1 eight:2 batch:5 alternative:1 robustness:1 thomas:1 top:1 include:2 konda:2 build:1 approximating:1 objective:2 added:4 question:1 occurs:1 already:1 strategy:1 rt:5 usual:2 gradient:7 distance:1 separate:3 thank:1 separating:1 s_:8 reason:1 length:2 retained:2 minimizing:1 difficult:1 unfortunately:2 executed:4 esults:1 subproblems:1 policy:29 twenty:1 perform:4 observation:1 revised:1 markov:1 benchmark:2 finite:4 rdf:6 descent:2 acknowledge:1 beat:1 immediate:1 mansour:2 discovered:1 arbitrary:1 pair:3 orst:2 learned:2 nip:6 address:2 able:1 departure:7 reading:1 max:1 reliable:1 suitable:1 difficulty:1 boyan:6 rely:1 residual:1 representing:1 shop:6 improve:3 mdps:4 imply:1 f4962098:1 axis:3 coupled:1 pleased:1 relative:1 afosr:1 acyclic:1 approximator:8 versus:1 validation:2 last:1 tsitsiklis:2 bias:1 sparse:1 tolerance:2 distributed:1 boundary:5 world:8 avoids:1 computes:1 stuck:1 made:1 reinforcement:14 replicated:1 author:1 san:4 approximate:1 preferred:1 global:4 investigating:1 summing:1 a_:3 francisco:4 search:6 continuous:1 table:8 terminate:1 robust:1 ca:4 improving:1 necessarily:1 domain:8 s2:1 child:6 repeated:1 referred:1 fashion:1 slow:1 scheduler:1 lie:1 theorem:3 down:1 explored:4 experimented:2 essential:1 incorporating:1 importance:1 generalizing:1 led:1 simply:1 failed:1 monotonic:1 applies:1 aa:11 ma:1 succeed:1 goal:1 consequently:1 careful:1 replace:2 experimentally:1 except:1 conservative:1 called:1 experimental:1 xin:1 e:1 est:2 bisector:1 puddle:4 internal:1 support:8 incorporate:1 tested:2
1,116
2,018
The Fidelity of Local Ordinal Encoding Javid Sadr, Sayan Mukherjee, Keith Thoresz, Pawan Sinha Center for Biological and Computational Learning Department of Brain and Cognitive Sciences, MIT Cambridge, Massachusetts, 02142 USA {sadr,sayan,thorek,sinha}@ai.mit.edu Abstract A key question in neuroscience is how to encode sensory stimuli such as images and sounds. Motivated by studies of response properties of neurons in the early cortical areas, we propose an encoding scheme that dispenses with absolute measures of signal intensity or contrast and uses, instead, only local ordinal measures. In this scheme, the structure of a signal is represented by a set of equalities and inequalities across adjacent regions. In this paper, we focus on characterizing the fidelity of this representation strategy. We develop a regularization approach for image reconstruction from ordinal measures and thereby demonstrate that the ordinal representation scheme can faithfully encode signal structure. We also present a neurally plausible implementation of this computation that uses only local update rules. The results highlight the robustness and generalization ability of local ordinal encodings for the task of pattern classification. 1 Introduction Biological and artificial recognition systems face the challenge of grouping together differing proximal stimuli arising from the same underlying object. How well the system succeeds in overcoming this challenge is critically dependent on the nature of the internal representations against which the observed inputs are matched. The representation schemes should be capable of efficiently encoding object concepts while being tolerant to their appearance variations. In this paper, we introduce and characterize a biologically plausible representation scheme for encoding signal structure. The scheme employs a simple vocabulary of local ordinal relations, of the kind that early sensory neurons are capable of extracting. Our results so far suggest that this scheme possesses several desirable characteristics, including tolerance to object appearance variations, computational simplicity, and low memory requirements. We develop and demonstrate our ideas in the visual domain, but they are intended to be applicable to other sensory modalities as well. The starting point for our proposal lies in studies of the response properties of neurons in the early sensory cortical areas. These response properties constrain Figure 1: (a) A schematic contrast response curve for a primary visual cortex neuron. The response of the neuron saturates at low contrast values. (b) An idealization of (a). This unit can be thought of as an ordinal comparator, providing information only about contrast polarity but not its magnitude. the kinds of measurements that can plausibly be included in our representation scheme. In the visual domain, many striate cortical neurons have rapidly saturating contrast response functions [1, 4]. Their tendency to reach ceiling level responses at low contrast values render these neurons sensitive primarily to local ordinal, rather than metric, relations. We propose to use an idealization of such units as the basic vocabulary of our representation scheme (figure 1). In this scheme, objects are encoded as sets of local ordinal relations across image regions. As discussed below, this very simple idea seems well suited to handling the photometric appearance variations that real-world objects exhibit. Figure 2: The challenge for a representation scheme: to construct stable descriptions of objects despite radical changes in appearance. As figure 2 shows, variations in illumination significantly alter the individual brightness of different parts of the face, such as the eyes, cheeks, and forehead. Therefore, absolute image brightness distributions are unlikely to be adequate for classifying all of these images as depicting the same underlying object. Even the contrast magnitudes across different parts of the face change greatly under different lighting conditions. While the absolute luminance and contrast magnitude information is highly variable across these images, Thoresz and Sinha [9] have shown that one can identify some stable ordinal measurements. Figure 3 shows several pairs of average brightness values over localized patches for each of the three images included in figure 2. Certain regularities are apparent. For instance, the average brightness of the left eye is always less than that of the forehead, irrespective of the lighting conditions. The relative magnitudes of the two brightness values may change, but the sign of the inequality does not. In other words, the ordinal relationship between the average brightnesses of the <left-eye, forehead> pair is invariant under lighting changes. Figure 3 shows several other such pair-wise invariances. It seems, therefore that local ordinal relations may encode the stable facial attributes across different illumination conditions. An additional advantage to using ordinal relations is their natural robustness to sensor noise. Thus, it would seem that local ordinal representations may be well suited for devising compact representations, robust against Figure 3: The absolute brightnesses and their relative magnitudes change under different lighting conditions but several pair-wise ordinal relationships stay invariant. large photometric variations, for at least some classes of objects. Notably, for similar reasons, ordinal measures have also been shown to be a powerful tool for simple, efficient, and robust stereo image matching [3]. In what follows, we address an important open question regarding the expressiveness of the ordinal representation scheme. Given that this scheme ignores absolute luminance and contrast magnitude information, an obvious question that arises is whether such a crude representation strategy can encode object/image structure with any fidelity. 2 Information Content of Local Ordinal Encoding Figure 4 shows how we define ordinal relations between an image region pa and its immediate neighbors pb = {pa1 , . . . , pa8 }. In the conventional rectilinear grid, when all image regions pa are considered, four of the eight relations are redundant; we encode the remaining four as {1, 0, ?1} based on the difference in luminance between two neighbors being positive, zero, or negative, respectively. To demonstrate the richness of information encoded by this scheme, we compare the original image to one produced by a function that reconstructs the image using local ordinal relationships as constraints. Our reconstruction function has the form f (x) = w ? ?(x), (1) where x = {i, j} is the position of a pixel, f (x) is its intensity, ? is a map from the input space into a high (possibly infinite) dimensional space, w is a hyperplane in this high-dimensional space, and u ? v denotes an inner product. Infinitely many reconstruction functions could satisfy the given ordinal constraints. To make the problem well-posed we regularize [10] the reconstruction function subject to the ordinal constraints, as done in ordinal regression for ranking document Department of Brain Sciences, MIT Cambridge, Massachusetts, USA. {sadr,sayan,thorek,sinha}@ai.mit.edu Neighbors? relations to pixel of interest ???????????????????????????I(pa ) < = < < > < < < I(pa1 ) I(pa2 ) I(pa3 ) I(pa4 ) I(pa5 ) I(pa6 ) I(pa7 ) I(pa8 ) (1) Figure 4: Ordinal relationships between an image region pa and its neighbors. ???????????????????????????retrieval results [5]. Our regularization term is a norm in a Reproducing Kernel Hilbert Space (RKHS) [2, 11]. Minimizing the norm in a RKHS subject to the ordinal constraints corresponds to the following convex constrained quadratic optimization problem: X 1 min ||w||2 + C ?p (2) ?,w 2 p subject to ?(?p )w ? (?(xpa ) ? ?(xpb )) ? |?p | ? ?p , ? p and ? ? 0, (3) where the function ?(y) = +1 for y ? 0 and ?1 otherwise, p is the index over all pairwise ordinal relations between all pixels pa and their local neighbors pb (as depicted in figure 4), ?p are slack variables which are penalized by C (the trade-off between smoothness and ordinal constraints), and ?p take integer values {?1, 0, 1} denoting the ordinal relation (less than, equal to, or greater than, respectively) between pa and pb ; for the case ?p = 0 the inequality in (3) will be a strict equality. Taking the dual of (2) subject to constraints (3) results in the following convex quadratic optimization problem which has only box constraints: X 1 XX ? pq ?p ?q K (4) max |?p | ?p ? ? 2 p q p subject to 0 ? ?p ? C ?C ? ?p ? C ?C ? ?p ? 0 if if if ?p > 0, ?p = 0, ?p < 0, (5) ? have where ?p are the dual Lagrange multipliers, and the elements of the matrix K the form ? pq K = (?(xpa ) ? ?(xpb )) ? (?(xqa ) ? ?(xqb )) = K(xpa , xqa ) ? K(xpb , xqa ) ? K(xpa , xqb ) + K(xpb , xqb ), where K(y, x) = ?(y)??(x) using the standard kernel trick [8]. In this paper we use only Gaussian kernels K(y, x) = exp(?||x?y||2 /2? 2 ). The reconstruction function, f (x), obtained from optimizing (4) subject to box constraints (5) has the following form X f (x) = ?p (K(x, xpa ) ? K(x, xpb )) . (6) p Note that in general many of the ?p values may be zero ? these terms do not contribute to the reconstruction, and the corresponding constraints in (3) were not 300 200 100 0 0 128 255 128 255 300 200 100 0 (a) (b) (c) 0 (d) Figure 5: Reconstruction results from the regularization approach. (a) Original images. (b) Reconstructed images. (c) Absolute difference between original and reconstruction. (d) Histogram of absolute difference. required. The remaining ?p with absolute value less than C satisfy the inequality constraints in (3), whereas those with absolute value at C violate them. Figure 5 depicts two typical reconstructions performed by this algorithm. The difference images and error histograms suggests that the reconstructions closely match the source images. 3 Discussion Our reconstruction results suggest that the local ordinal representation can faithfully encode image structure. Thus, even though individual ordinal relations are insensitive to absolute luminance or contrast magnitude, a set of such relations implicitly encodes metric information. In the context of the human visual system, this result suggests that the rapidly saturating contrast response functions of the early visual neurons do not significantly hinder their ability to convey accurate image information to subsequent cortical stages. An important question that arises here is what are the strengths and limitations of local ordinal encoding. The first key limitation is that for any choice of neighborhood size over which ordinal relations are extracted, there are classes of images for which the local ordinal representation will be unable to encode the metric structure. For a neighborhood of size n, an image with regions of different luminance embedded in a uniform background and mutually separated by a distance greater than n would constitute such an image. In general, sparse images present a problem for this representation scheme, as might foveal or cortical ?magnification,? for example. This issue could be addressed by using ordinal relations across multiple scales, perhaps in an adaptive way that varies with the smoothness or sparseness of the stimulus. Second, the regularization approach above seems biologically implausible. Our intent in using this approach for reconstructions was to show via well-understood theoretical tools the richness of information that local ordinal representations pro- Figure 6: Reconstruction results from the relaxation approach. vide. In order to address the neural plausibility requirement, we have developed a simple relaxation-based approach with purely local update rules of the kind that can easily be implemented by cortical circuitry. Each unit communicates only with its immediate neighbors and modifies its value incrementally up or down (starting from an arbitrary state) depending on the number of ordinal relations in the positive or negative direction. This computation is performed iteratively until the network settles to an equilibrium state. The update rule can be formally stated as X (?(Rpa ,t ? Rpb ,t ) ? ?(Ipa ? Ipb )), (7) Rpa ,t+1 = Rpa ,t + ? pb where Rpa ,t is the intensity of the reconstructed pixel pa at step t, Ipa is the intensity of the corresponding pixel in the original image, ? is a positive update rate, and ? and pb are as described above. Figure 6 shows four examples of image reconstructions performed using a relaxation-based approach. A third potential limitation is that the scheme does not appear to constitute a compact code. If each pixel must be encoded in terms of its relations with all of its eight neighbors, where each relation takes one of three values, {?1, 0, 1}, then what has been gained over the original image where each pixel is encoded by 8 bits? There are three ways to address this question. 1. Eight relations per pixel is highly redundant ? four are sufficient. In fact, as shown in figure 7, the scheme can also tolerate several missing relations. Figure 7: Five reconstructions, shown here to demonstrate the robustness of local ordinal encoding to missing inputs. From left to right: reconstructions based on 100%, 80%, 60%, 40%, and 20% of the full set of immediate neighbor relations. 2. An advantage to using ordinal relations is that they can be extracted and transmitted much more reliably than metric ones. These relations share the same spirit (a) (b) Figure 8: A small collection of ordinal relations (a), though insufficient for high fidelity reconstruction, is very effective for pattern classification despite significant appearance variations. (b) Results of using a local ordinal relationship based template to detect face patterns. The program places white dots at the centers of patches classified as faces. (From Thoresz and Sinha, in preparation.) as loss functions used in robust statistics [6] and trimmed or Winsorized estimators. 3. The intent of the visual system is often not to encode/reconstruct images with perfect fidelity, but rather to encode the most stable characteristics that can aid in classification. In this context, a few ordinal relations may suffice for encoding objects reliably. Figure 8 shows the results of using less than 20 relations for detecting faces. Clearly, such a small set would not be sufficient for reconstructions, but it works well for classification. Its generalization arises because it defines an equivalence class of patterns. In summary, the ordinal representation scheme provides a neurally plausible strategy for encoding signal structure. While in this paper we focus on demonstrating the fidelity of this scheme, we believe that its true strength lies in defining equivalence classes of patterns enabling generalizations over appearance variations in objects. Several interesting directions remain to be explored. These include the study of ordinal representations across multiple scales, learning schemes for identifying subsets of ordinal relations consistent across different instances of an object, and the relationship of this work to multi-dimensional scaling [12] and to the use of truncated, quantized wavelet coefficients as ?signatures? for fast, multiresolution image querying [7]. Acknowledgements We would like to thank Gadi Geiger, Antonio Torralba, Ryan Rifkin, Gonzalo Ramos, and Tabitha Spagnolo. Javid Sadr is a Howard Hughes Medical Institute Pre-Doctoral Fellow. References [1] A. Anzai, M. A. Bearse, R. D. Freeman, and D. Cai. Contrast coding by cells in the cat?s striate cortex: monocular vs. binocular detection. Visual Neuroscience, 12:77?93, 1995. [2] N. Aronszajn. Theory of reproducing kernels. Trans. Amer. Math. Soc., 686:337?404, 1950. [3] D. Bhat and S. Nayar. Ordinal measures for image correspondence. In IEEE Conf. on Computer Vision and Pattern Recognition, pages 351?357, 1996. [4] G. C. DeAngelis, I. Ohzawa, and R. D. Freeman. Spatiotemporal organization of simple-cell receptive fields in the cat?s striate cortex. i. general characteristics and postnatal development. J. Neurophysiology, 69:1091?1117, 1993. [5] R. Herbrich, T. Graepel, and K. Obermeyer. Support vector learning for ordinal regression. In Proc. of the Ninth Intl. Conf. on Artificial Neural Networks, pages 97?102, 1999. [6] P. Huber. Robust Statistics. John Wiley and Sons, New York, 1981. [7] C. E. Jacobs, A. Finkelstein, and D. H. Salesin. Fast multiresolution image querying. In Computer Graphics Proc., Annual Conf. Series (SIGGRAPH 95), pages 277?286, 1995. [8] T. Poggio. On optimal nonlinear associative recall. Biological Cybernetics, 19:201?209, 1975. [9] K. Thoresz and P. Sinha. Qualitative representations for recognition. Vision Sciences Society Abstracts, 1:81, 2001. [10] A. N. Tikhonov and V. Y. Arsenin. Solutions of Ill-posed Problems. W. H. Winston, Washington, D.C., 1977. [11] G. Wahba. Spline Models for Observational Data. Series in Applied Mathematics, Vol 59, SIAM, Philadelphia, 1990. [12] F. W. Young and C. H. Null. Mds of nominal data: the recovery of metric information with alscal. Psychometika, 53.3:367?379, 1978.
2018 |@word neurophysiology:1 seems:3 norm:2 open:1 jacob:1 brightness:7 thereby:1 foveal:1 series:2 denoting:1 document:1 rkhs:2 must:1 john:1 subsequent:1 update:4 v:1 devising:1 postnatal:1 provides:1 detecting:1 contribute:1 quantized:1 math:1 herbrich:1 five:1 ipb:1 qualitative:1 vide:1 introduce:1 pairwise:1 huber:1 notably:1 multi:1 brain:2 freeman:2 xx:1 underlying:2 matched:1 suffice:1 null:1 what:3 kind:3 developed:1 differing:1 fellow:1 unit:3 medical:1 appear:1 positive:3 understood:1 local:19 despite:2 encoding:10 might:1 doctoral:1 equivalence:2 suggests:2 hughes:1 area:2 thought:1 significantly:2 matching:1 word:1 pre:1 suggest:2 context:2 conventional:1 map:1 center:2 missing:2 modifies:1 starting:2 convex:2 simplicity:1 identifying:1 recovery:1 rule:3 estimator:1 regularize:1 variation:7 nominal:1 us:2 pa:7 element:1 trick:1 recognition:3 magnification:1 mukherjee:1 observed:1 region:6 richness:2 trade:1 hinder:1 signature:1 purely:1 easily:1 siggraph:1 represented:1 cat:2 separated:1 alscal:1 fast:2 effective:1 deangelis:1 artificial:2 neighborhood:2 apparent:1 encoded:4 posed:2 plausible:3 pa2:1 otherwise:1 reconstruct:1 ability:2 statistic:2 associative:1 advantage:2 cai:1 propose:2 reconstruction:18 product:1 finkelstein:1 rapidly:2 rifkin:1 multiresolution:2 description:1 regularity:1 requirement:2 intl:1 perfect:1 object:12 depending:1 develop:2 radical:1 keith:1 soc:1 implemented:1 direction:2 closely:1 attribute:1 human:1 settle:1 observational:1 anzai:1 generalization:3 biological:3 ryan:1 considered:1 exp:1 equilibrium:1 circuitry:1 early:4 torralba:1 proc:2 applicable:1 sensitive:1 xpa:5 faithfully:2 tool:2 mit:4 clearly:1 sensor:1 always:1 gaussian:1 rather:2 encode:9 focus:2 greatly:1 contrast:12 detect:1 dependent:1 unlikely:1 relation:26 rpa:4 pixel:8 issue:1 fidelity:6 classification:4 dual:2 ill:1 development:1 constrained:1 equal:1 construct:1 field:1 washington:1 alter:1 photometric:2 stimulus:3 spline:1 employ:1 primarily:1 few:1 individual:2 intended:1 pawan:1 detection:1 organization:1 interest:1 highly:2 gadi:1 accurate:1 capable:2 poggio:1 facial:1 theoretical:1 sinha:6 instance:2 subset:1 uniform:1 graphic:1 characterize:1 varies:1 spatiotemporal:1 proximal:1 siam:1 stay:1 off:1 together:1 reconstructs:1 possibly:1 cognitive:1 conf:3 potential:1 coding:1 coefficient:1 satisfy:2 ranking:1 performed:3 characteristic:3 efficiently:1 identify:1 salesin:1 critically:1 produced:1 lighting:4 cybernetics:1 classified:1 implausible:1 reach:1 against:2 obvious:1 massachusetts:2 recall:1 hilbert:1 graepel:1 tolerate:1 response:8 amer:1 done:1 box:2 though:2 stage:1 binocular:1 until:1 aronszajn:1 nonlinear:1 incrementally:1 defines:1 perhaps:1 believe:1 usa:2 ohzawa:1 concept:1 multiplier:1 true:1 equality:2 regularization:4 iteratively:1 white:1 adjacent:1 demonstrate:4 pro:1 image:31 wise:2 insensitive:1 discussed:1 forehead:3 measurement:2 significant:1 cambridge:2 ai:2 rectilinear:1 smoothness:2 grid:1 mathematics:1 gonzalo:1 pq:2 dot:1 stable:4 cortex:3 optimizing:1 tikhonov:1 certain:1 inequality:4 transmitted:1 additional:1 greater:2 redundant:2 signal:5 neurally:2 desirable:1 violate:1 sound:1 multiple:2 full:1 match:1 plausibility:1 retrieval:1 schematic:1 basic:1 regression:2 vision:2 metric:5 histogram:2 kernel:4 cell:2 proposal:1 background:1 whereas:1 bhat:1 addressed:1 source:1 modality:1 posse:1 strict:1 subject:6 spirit:1 seem:1 integer:1 extracting:1 wahba:1 inner:1 idea:2 regarding:1 whether:1 motivated:1 trimmed:1 stereo:1 render:1 york:1 constitute:2 adequate:1 antonio:1 ipa:2 sign:1 neuroscience:2 arising:1 per:1 vol:1 key:2 four:4 demonstrating:1 pb:5 luminance:5 relaxation:3 idealization:2 powerful:1 place:1 pa1:2 cheek:1 patch:2 geiger:1 scaling:1 bit:1 correspondence:1 quadratic:2 winston:1 annual:1 strength:2 constraint:10 constrain:1 encodes:1 min:1 department:2 across:8 remain:1 son:1 biologically:2 invariant:2 handling:1 ceiling:1 monocular:1 mutually:1 slack:1 ordinal:46 eight:3 robustness:3 original:5 denotes:1 remaining:2 include:1 plausibly:1 society:1 question:5 strategy:3 primary:1 receptive:1 striate:3 md:1 exhibit:1 distance:1 unable:1 thank:1 reason:1 code:1 index:1 polarity:1 relationship:6 providing:1 minimizing:1 insufficient:1 negative:2 stated:1 intent:2 implementation:1 reliably:2 neuron:8 howard:1 enabling:1 truncated:1 immediate:3 defining:1 saturates:1 reproducing:2 ninth:1 arbitrary:1 intensity:4 expressiveness:1 overcoming:1 pair:4 required:1 trans:1 address:3 below:1 pattern:6 challenge:3 program:1 including:1 memory:1 max:1 natural:1 ramos:1 scheme:20 eye:3 irrespective:1 philadelphia:1 acknowledgement:1 relative:2 embedded:1 loss:1 highlight:1 interesting:1 limitation:3 querying:2 localized:1 sufficient:2 consistent:1 classifying:1 share:1 arsenin:1 penalized:1 summary:1 institute:1 neighbor:8 template:1 characterizing:1 face:6 taking:1 absolute:10 sparse:1 tolerance:1 curve:1 cortical:6 vocabulary:2 world:1 sensory:4 ignores:1 collection:1 adaptive:1 far:1 reconstructed:2 compact:2 implicitly:1 tolerant:1 winsorized:1 nature:1 robust:4 dispenses:1 depicting:1 domain:2 noise:1 convey:1 depicts:1 aid:1 wiley:1 position:1 lie:2 crude:1 communicates:1 third:1 wavelet:1 young:1 sayan:3 down:1 explored:1 grouping:1 gained:1 magnitude:7 illumination:2 sparseness:1 suited:2 depicted:1 appearance:6 infinitely:1 visual:7 lagrange:1 saturating:2 corresponds:1 extracted:2 comparator:1 content:1 change:5 included:2 infinite:1 typical:1 hyperplane:1 invariance:1 tendency:1 succeeds:1 formally:1 internal:1 support:1 arises:3 preparation:1 nayar:1
1,117
2,019
Learning Body Pose via Specialized Maps Romer Rosales Department of Computer Science Boston University, Boston, MA 02215 rrosales@cs.bu.edu Stan Sclaroff Department of Computer Science Boston University, Boston, MA 02215 sclaroff@cs.bu.edu Abstract A nonlinear supervised learning model, the Specialized Mappings Architecture (SMA), is described and applied to the estimation of human body pose from monocular images. The SMA consists of several specialized forward mapping functions and an inverse mapping function. Each specialized function maps certain domains of the input space (image features) onto the output space (body pose parameters). The key algorithmic problems faced are those of learning the specialized domains and mapping functions in an optimal way, as well as performing inference given inputs and knowledge of the inverse function. Solutions to these problems employ the EM algorithm and alternating choices of conditional independence assumptions. Performance of the approach is evaluated with synthetic and real video sequences of human motion. 1 Introduction In everyday life, humans can easily estimate body part locations (body pose) from relatively low-resolution images of the projected 3D world (e.g., when viewing a photograph or a video). However, body pose estimation is a very difficult computer vision problem. It is believed that humans employ extensive prior knowledge about human body structure and motion in this task [10]. Assuming this , we consider how a computer might learn the underlying structure and thereby infer body pose. In computer vision, this task is usually posed as a tracking problem. Typically, models comprised of 2D or 3D geometric primitives are designed for tracking a specific articulated body [13, 5, 2, 15]. At each frame, these models are fitted to the image to optimize some cost function. Careful manual placement of the model on the first frame is required, and tracking in subsequent frames tends to be sensitive to errors in initialization and numerical drift. Generally, these systems cannot recover from tracking errors in the middle of a sequence. To address these weaknesses, more complex dynamic models have been proposed [14, 13,9]; these methods learn a prior over some specific motion (such as walking). This strong prior however, substantially limits the generality of the motions that can be tracked. Departing from the aforementioned tracking paradigm, in [8] a Gaussian probability model was learned for short human motion sequences. In [17] dynamic programming was used to calculate the best global labeling according to the learned joint probability density function of the position and velocity of body features. Still, in these approaches, the joint locations, correspondences, or model initialization must be provided by hand. In [1], the manifold of human body dynamics was modeled via a hidden Markov model and learned via entropic minimization. In all of these approaches models were learned. Although the approach presented here can be used to model dynamics, we argue that when general human motion dynamics are intended to be learned, the amount of training data, model complexity, and computational resources required are impractical. As a consequence, models with large priors towards specific motions (e .g., walking) are generated. In this paper we describe a non-linear supervised learning algorithm, the Specialized Maps Architecture (SMA), for recovering articulated body pose from single monocular images. This approach avoids the need for initialization and tracking per se, and reduces the above mentioned disadvantages. 2 Specialized Maps There at least two key characteristics of the problem we are trying to solve which make it different from other supervised learning problems. First, we have access to the inverse map. We are trying to learn unknown probabilistic maps from inputs to outputs space, but we have access to the map (in general probabilistic) from outputs to inputs. In our pose estimation problem, it is easy to see how we can artificially, using computer graphics (CG), produce some visual features (e.g., body silhouettes) given joint positions 1 . Second, it is one-to-many: one input can be associated with more than one output. Features obtained from silhouettes (and many other visual features) are ambiguous. Consider an occluded arm, or the reflective ambiguity generated by symmetric poses. This last observation precludes the use of standard algorithms for supervised learning that fit a single mapping function to the data. Given input and output spaces ~c and ~t, and the inverse function ( : ~t -+ ~c, we describe a solution for these supervised learning problems. Our approach consists in generating a series of m functions ?k : ~c -+ ~t. Each of these functions is specialized to map only certain inputs (for a specialized sub-domain) better than others. For example, each sub-domain can be a region of the input space. However, the specialized sub-domain of ?k can be more general than just a connected region in the input space. Several other learning models use a similar concept of fitting surfaces to the observed data by splitting the input space into several regions and approximating simpler functions in these regions (e.g., [11,7, 6]). However, in these approaches, the inverse map is not incorporated in the estimation algorithm because it is not considered in the problem definition and the forward model is usually more complex, making inference and learning more difficult. The key algorithmic problems are that of estimating the specialized domains and functions in an optimal way (taking into account the form of the specialized functions), and using the knowledge of the inverse function to formulate efficient inferIThus, ( is a computer graphics rendering, in general called forward kinematics ence and learning algorithms. We propose to determine the specialized domains and functions using an approximate EM algorithm and to perform inference using, in an alternating fashion, the conditional independence assumptions specified by the forward and inverse models. Fig. l(a) illustrates a learned forward model. Figure 1: SMA diagram illustrating (a) an already learned SMA model with m specialized functions mapping subsets of the training data, each subset is drawn with a different color (at initializations, coloring is random) and (b) the mean-output inference process in which a given observation is mapped by all the specialized functions , and then a feedback matching step, using (, is performed to choose the best of the m estimates. 3 Probabilistic Model Let the training sets of output-input observations be \)! = {1jI1, ... , 1jIN } , and Y = {Vl , ... ,VN} respectively. We will use Z i = (1jIi,Vi) to define the given output-input training pair, and Z = {ZI ' ... , ZN } as our observed training set. We introduce the unobserved random variable y = (Yl , ... , Yn). In our model any Yi has domain the discrete set C = {l, ... , M} oflabels for the specialized functions , and can be thought as the function number used to map data point i; thus M is the number of specialized mapping functions. Our model uses parameters 8 = (8 1 , ... , 8M , A) , 8k represents the parameters of the mapping function k; A = (AI"", AM), where Ak represents P(Yi = kI8): the prior probability that mapping function with label i will be used to map an unknown point. As an example, P(Yi lz i, 8) represents the probability that function number Yi generated data point number i. Using Bayes' rule and assuming independence of observations given 8, we have the log-probability of our data given the modellogp(ZI8), which we want to maximize: argm;x 2:)og LP(1jIi lvi, Yi = k,8)P(Yi = kI8)p(Vi ), i (1) k where we used the independence assumption p(vI8) = p(v). This is also equivalent to maximizing the conditional likelihood of the model. Because of the log-sum encountered, this problem is intractable in general. However, there exist practical approximate optimization procedures, one of them is Expectation Maximization (EM) [3,4, 12]. 3.1 Learning The EM algorithm is well known, therefore here we only provide the derivations specific to SMA's. The E-step consists of finding P(y = klz, 8) = P(y). Note that the variables Yi are assumed independent (given Z i)' Thus, factorizing P(y): p(y) = II P(t)(Yi) = II[(AYiP(1/Jilvi,Yi,B))/(2:AkP(1/Jilvi,Yi = k,B))] (2) kEC However, p( 1/Ji lVi, Yi = k, B) is still undefined. For the implementation described in this paper we use N(1/Ji; ?k(Vi,B k ), ~k)' where Bk are the parameters of the k-th specialized function, and ~k the error covariance of the specialized function k . One way to interpret this choice is to think that the error cost in estimating 1/J once we know the specialized function to use, is a Gaussian distribution with mean the output of the specialized function and some covariance which is map dependent. This also led to tractable further derivations. Other choices were given in [16]. The M-step consists of finding B(t) = argmaxoEj>(t) [logp(Z,y IB)]. In our case we can show that this is equivalent to finding: argmJn 2: 2: P(t)(Yi = k)(1/Ji i ?k(Vi, Bk))T~kl(Zi - ?k(Zi,B k ))? (3) k This gives the following update rules for Ak and ~k (where Lagrange multipliers were used to incorporate the constraint that the sum of the Ak'S is 1. -n1 2:. P(Yi = klzi' B) (4) In keeping the formulation general, we have not defined the form of the specialized functions ?k. Whether or not we can find a closed form solution for the update of Bk depends on the form of ?k. For example if ?k is a non-linear function, we may have to use iterative optimization to find Bit). In case ?k yield a quadratic form, then a closed form update exists. However, in general we have: (6) In our experiments, ?k is a I-hidden layer perceptron. Thus, the M-step is an approximate, iterative optimization procedure. 4 Inference Once learning is accomplished, each specialized function maps (with different levels of accuracy) the input space. We can formally state the inference process as that of maximum-a-posteriori (MAP) estimation where we are interested in finding the most likely output h given an input configuration x: h* = argmaxp(hlx) = argmax ' " p(hly, x)P(y), h h ~ (7) Y Any further treatment depends on the properties of the probability distributions involved. If p(hlx, y) = N(h ; ?y(x) , ~y), the MAP estimate involves finding the maximum in a mixture of Gaussians. However, no closed form solution exists and moreover, we have not incorporated the potentially useful knowledge of the inverse function C. 4.1 MAP by Using the Inverse Function ( The access to a forward kinematics function ( (called here the inverse function) allows to formulate a different inference algorithm. We are again interested in finding an optimal h* given an input x (e.g. , an optimal body pose given features taken from an image). This can be formulated as: (8) h* = arg maxp(hlx) = argmaxp(xlh) "p(hly, x)P(y) , h ~ h y simply by Bayes' rule, and marginalizing over all variables except h. Note that we have made the distribution p(xlh) appear in the solution. This is important because we can know use our knowledge of ( to define this distribution. This solution is completely general within our architecture, we did not make any assumptions on the form of the distributions or algorithms used. 5 Approximate Inference using ( Let us assume that we can approximate Lyp(hly, x)P(y) by a set of samples generated according to p(hly,x)P(y) and a kernel function K(h,hs). Denote the set of samples HSpl = {hs}s=l...s. An approximate to Lyp(hly,x)P(y) is formally built by ~ L;=l K(h , h s ), with the normalizing condition any given h s . J K(h , hs)dh = 1 for We will consider two simple forms of K. If K(h, h s ) = J(h - h s ), we have: argmaxhP(xlh) L;=l J(h - h s). h= After some simple manipulations, this can be reduced to the following equivalent discrete optimization problem whose goal is to find the most likely sample s*: (9) where the last equivalence used the assumption p(xlh) = N(x; ((h), ~d. A S If K(h, h s) = N(h ; hs , ~Spl)' we have: h = argmaxhP(xlh) L S =l N(h ; hs , ~Spl). This case is hard to use in practice, because contrary to the case above (Eq. 9) , in general, there is no guarantee that the optimal h is among the samples. 5.1 A Deterministic Approximation based on the Functions Mean Output The structure of the inference in SMA, and the choice of probabilities p(hlx, y) allows us to construct a newer approximation that is considerably less expensive to compute, and it is deterministic. Intuitively they idea consists of asking each of the specialized functions ?k what their most likely estimate for h is, given the observed input x. The opinions of each of these specialized functions are then evaluated using our distribution p(xlh) similar to the above sampling method. This can be justified by the observation that the probability of the mean is maximal in a Gaussian distribution. Thus by considering the means ?k(X), we would be considering the most likely output of each specialized function. Of course, in many cases this approximation could be very far from the best solution, for example when the uncertainty in the function estimate is relatively high relative to the difference between means. We use Fig. l(b) to illustrate the mean-output (MO) approximate inference process. When generating an estimate of body pose, denoted h, given an input x (the gray point with a dark contour in the lower plane), the SMA generates a series of output hypotheses tl q, = {h!h obtained using hk = (/Jk(x), with k E C (illustrated by each of the points pointed to by the arrows). Given the set tlq" the most accurate hypothesis under the mean-output criteria is the one that minimizes the function: k* (10) where in the last equation we have assumed p(xlh) is Gaussian. 5.2 Bayesian Inference Note that in many cases, there may not be any need to simply provide a point estimate, in terms of a most likely output h. In fact we could instead use the whole distribution found in the inference process. We can show that using the above choices for K we can respectively obtain. 1 s p(hlx) = S 2: N (x; ((hs ), ~d, (11) 8= 1 s p(hlx) = N(h; h8' ~Spz) 2:N(x; ((h) , ~d? (12) 8=1 6 Experiments The described architecture was tested using a computer graphics rendering as our ( inverse function. The training data set consisted of approx. 7,000 frames of human body poses obtained through motion capture. The output consisted of 20 2D marker positions (i. e., 3D markers projected to the image plane using a perspective model) but linearly encoded by 8 real values using Principal Component Analysis (PCA). The input (visual features) consisted of 7 real-valued Hu moments computed on synthetically generated silhouettes of the articulated figure. For training/testing we generated 120,000 data points: our 3D poses from motion capture were projected to 16 views along the view-sphere equator. We took 8,000 for training and the rest for testing. The only free parameter in this test, related to the given SMA, was the number of specialized functions used; this was set to 15. For this, several model selection approaches could be used instead. Due to space limitations, in this paper we show results using the mean-output inference algorithm only, readers are referred to http://cs-people.bu.edu/rrosales/SMABodyInference where inference using multiple samples is shown. Fig. 2(left) shows the reconstruction obtained in several single images coming from three different artificial sequences. The agreement between reconstruction and observation is easy to perceive for all sequences. Note that for self-occluding configurations, reconstruction is harder, but still the estimate is close to ground-truth. No human intervention nor pose initialization was required. For quantitative results, Fig. 2(right) shows the average marker error and variance per body orientation in percentage of body height. Note that the error is bigger for orientations closer to a and 7r radians. This intuitively agrees with the notion that at those angles (side-views) , there is less visibility of the body parts. We consider this performance promising, given the complexity of the task and the simplicity of the approach. By choosing poses at random from training set, the RMSE was 17% of body height. In related work, quantitative performance have been usually ignored, in part due to the lack of ground-truth and standard evaluation data sets. Penormance regarding cameraviewpoinl (16 101al) 2.9 ,-----~~--,-:.----'.----.--:----.;...-_---,----, 2.75 14 16 Figure 2: Left: Example reconstruction of several test sequences with CG-generated silhouettes. Each set consists of input images and reconstruction (every 5th frame). Right: Marker root-mean-square-error and variance per camera viewpoint (every 27r/32 rads.). Units are percentage of body height. Approx. 110,000 test poses were used. 6.1 Experiments using Real Visual Cues Fig. 3 shows examples of system performance with real segmented visual data, obtained from observing a human subject. Reconstruction for several relatively complex sequences are shown. Note that even though the characteristics of the segmented body differ from the ones used for training, good performance is still achieved. Most reconstructions are visually close to what can be thought as the right pose reconstruction. Body orientation is also generally accurate. 7 Conclusion In this paper, we have proposed the Specialized Mappings Architecture (SMA) . A learning algorithm was developed for this architecture using ideas from ML estimation and latent variable models. Inference was based on the possibility of alternatively use different sets of conditional independence assumptions specified by the forward and inverse models. The incorporation of the inverse function in the model allows for simpler forward models. For example the inverse function is an architectural alternative to the gating networks of Mixture of Experts [11]. SMA advantages for body pose estimation include: no iterative methods for inference are used, the Figure 3: Reconstruction obtained from observing a human subject (every 10th frame). algorithm for inference runs in constant time and scales only linearly O(M) with respect to the number of specialized functions M; manual initialization is not required; compared to approaches that learn dynamical models, the requirements for data are much smaller, and also large priors to specific motions are prevented thus improving generalization capabilities. References [1] M. Brand. Shadow puppetry. In ICCV, 1999. [2] C. Bregler. Tracking people with twists and exponential maps. In CVPR, 1998. [3] 1. Csiszar and G. Thsnady. Information geometry and alternating minimization procedures. Statistics and Decisions, 1:205- 237, 1984. [4] A. Dempster, N. Laird, and D. Rubin. Maximum likelihood estimation from incomplete data. Journal of the Royal Statistical Society (B), 39(1), 1977. [5] J. Deutscher, A. Blake, and 1. Reid. Articulated body motion capture by annealed particle filtering. In CVPR, 2000. [6] J.H. Friedman. Multivatiate adaptive regression splines. The Annals of Statistics, 19,1-141 , 1991. [7] G. Hinton, B. Sallans, and Z. Ghahramani. A hierarchical community of experts. Learning in Graphical Models, M. Jordan (editor) , 1998. [8] N. Howe, M. Leventon, and B. Freeman. Bayesian reconstruction of 3d human motion from single-camera video. In NIPS-1 2, 2000. [9] M. Isard and A. Blake. Contour tracking by stochastic propagation of conditional density. In ECCV, 1996. [10] G. Johansson. Visual perception of biological motion and a model for its analysis. P erception and Psychophysics, 14(2): 210-211, 1973. [11] M. 1. Jordan and R. A. Jacobs. Hierarchical mixtures of experts and the EM algorithm. N eural Computation, 6, 181-214, 1994. [12] R. Neal and G. Hinton. A view of the em algorithm that justifies incremental , sparse, and other variants. Learning in Graphical Models, M. Jordan (editor) , 1998. [13] Dirk Ormoneit , Hedvig Sidenbladh, Michael J . Black, and Trevor Hastie. Learning and tracking cyclic human motion. In NIPS-1 3, 200l. [14] Vladimir Pavlovic, James M. Rehg, and John MacCormick. Learning switching linear models of human motion. In NIPS-13, 200l. [15] J. M. Regh and T. Kanade. Model-based tracking of self-occluding articulated objects. In ICC V, 1995. [16] R. Rosales and S. Sclaroff. Specialized mappings and the estimation of body pose from a single image. In IEEE Human Motion Workshop , 2000. [17] Y. Song, Xiaoling Feng, and P. Perona. Towards detection of human motion. In CVPR, 2000.
2019 |@word h:6 illustrating:1 middle:1 johansson:1 hu:1 covariance:2 jacob:1 thereby:1 harder:1 moment:1 configuration:2 series:2 cyclic:1 must:1 john:1 subsequent:1 numerical:1 visibility:1 designed:1 update:3 cue:1 isard:1 plane:2 short:1 argm:1 location:2 simpler:2 height:3 along:1 consists:6 fitting:1 introduce:1 nor:1 freeman:1 considering:2 provided:1 estimating:2 underlying:1 moreover:1 what:2 substantially:1 minimizes:1 developed:1 unobserved:1 finding:6 impractical:1 guarantee:1 quantitative:2 every:3 unit:1 intervention:1 yn:1 appear:1 reid:1 tends:1 limit:1 consequence:1 switching:1 puppetry:1 ak:3 might:1 black:1 initialization:6 equivalence:1 practical:1 camera:2 testing:2 practice:1 procedure:3 thought:2 matching:1 onto:1 cannot:1 selection:1 close:2 h8:1 optimize:1 equivalent:3 map:17 deterministic:2 maximizing:1 annealed:1 primitive:1 resolution:1 formulate:2 simplicity:1 splitting:1 perceive:1 rule:3 rehg:1 notion:1 annals:1 programming:1 us:1 hypothesis:2 agreement:1 velocity:1 expensive:1 jk:1 walking:2 observed:3 capture:3 calculate:1 region:4 connected:1 mentioned:1 dempster:1 complexity:2 occluded:1 dynamic:5 completely:1 easily:1 joint:3 derivation:2 articulated:5 describe:2 artificial:1 labeling:1 choosing:1 whose:1 encoded:1 posed:1 solve:1 valued:1 cvpr:3 precludes:1 maxp:1 statistic:2 think:1 laird:1 sequence:7 advantage:1 took:1 propose:1 reconstruction:10 maximal:1 coming:1 everyday:1 requirement:1 produce:1 generating:2 argmaxp:2 incremental:1 object:1 illustrate:1 pose:19 eq:1 strong:1 recovering:1 c:3 involves:1 shadow:1 rosales:2 differ:1 stochastic:1 human:17 viewing:1 opinion:1 generalization:1 biological:1 bregler:1 considered:1 ground:2 blake:2 visually:1 mapping:11 algorithmic:2 mo:1 sma:11 entropic:1 estimation:9 label:1 sensitive:1 agrees:1 minimization:2 gaussian:4 og:1 likelihood:2 hk:1 cg:2 am:1 posteriori:1 inference:17 dependent:1 vl:1 typically:1 hidden:2 perona:1 interested:2 arg:1 aforementioned:1 among:1 orientation:3 denoted:1 psychophysics:1 once:2 construct:1 sampling:1 represents:3 others:1 spline:1 hedvig:1 pavlovic:1 employ:2 intended:1 argmax:1 geometry:1 n1:1 friedman:1 detection:1 possibility:1 evaluation:1 weakness:1 mixture:3 undefined:1 csiszar:1 accurate:2 closer:1 incomplete:1 fitted:1 asking:1 ence:1 disadvantage:1 leventon:1 zn:1 logp:1 maximization:1 cost:2 subset:2 comprised:1 graphic:3 synthetic:1 considerably:1 density:2 bu:3 probabilistic:3 yl:1 michael:1 again:1 ambiguity:1 choose:1 xlh:7 expert:3 account:1 jii:2 vi:4 depends:2 performed:1 view:4 root:1 closed:3 observing:2 recover:1 bayes:2 capability:1 hly:5 rmse:1 square:1 accuracy:1 variance:2 characteristic:2 yield:1 ji1:1 bayesian:2 manual:2 trevor:1 definition:1 involved:1 james:1 associated:1 radian:1 treatment:1 knowledge:5 color:1 coloring:1 supervised:5 formulation:1 evaluated:2 though:1 generality:1 just:1 hand:1 hlx:6 nonlinear:1 marker:4 lack:1 propagation:1 gray:1 concept:1 multiplier:1 consisted:3 alternating:3 symmetric:1 neal:1 illustrated:1 self:2 ambiguous:1 criterion:1 argmjn:1 trying:2 motion:17 image:10 specialized:30 ji:3 tracked:1 twist:1 interpret:1 ai:1 approx:2 pointed:1 particle:1 access:3 surface:1 perspective:1 manipulation:1 certain:2 life:1 yi:13 accomplished:1 determine:1 paradigm:1 maximize:1 ii:2 multiple:1 infer:1 reduces:1 segmented:2 believed:1 sphere:1 prevented:1 bigger:1 variant:1 regression:1 vision:2 expectation:1 kernel:1 achieved:1 equator:1 justified:1 want:1 diagram:1 howe:1 rest:1 subject:2 contrary:1 jordan:3 reflective:1 synthetically:1 easy:2 rendering:2 independence:5 fit:1 zi:3 architecture:6 hastie:1 idea:2 regarding:1 whether:1 pca:1 song:1 ignored:1 generally:2 useful:1 se:1 amount:1 dark:1 reduced:1 http:1 exist:1 percentage:2 per:3 discrete:2 key:3 drawn:1 sum:2 run:1 inverse:14 akp:1 uncertainty:1 angle:1 reader:1 spl:2 architectural:1 vn:1 sallans:1 decision:1 bit:1 layer:1 correspondence:1 quadratic:1 encountered:1 placement:1 constraint:1 incorporation:1 generates:1 erception:1 performing:1 relatively:3 deutscher:1 department:2 according:2 smaller:1 em:6 newer:1 lp:1 making:1 intuitively:2 iccv:1 taken:1 monocular:2 resource:1 equation:1 kinematics:2 know:2 tractable:1 gaussians:1 hierarchical:2 romer:1 alternative:1 include:1 graphical:2 ghahramani:1 approximating:1 society:1 lyp:2 feng:1 already:1 sidenbladh:1 maccormick:1 mapped:1 manifold:1 argue:1 assuming:2 modeled:1 vladimir:1 difficult:2 potentially:1 implementation:1 unknown:2 perform:1 observation:6 markov:1 jin:1 hinton:2 incorporated:2 dirk:1 frame:6 community:1 drift:1 bk:3 pair:1 required:4 specified:2 extensive:1 kl:1 rad:1 learned:7 nip:3 address:1 usually:3 dynamical:1 perception:1 built:1 royal:1 video:3 ormoneit:1 arm:1 stan:1 faced:1 prior:6 geometric:1 icc:1 marginalizing:1 relative:1 limitation:1 filtering:1 rubin:1 viewpoint:1 editor:2 eccv:1 course:1 last:3 keeping:1 free:1 side:1 perceptron:1 taking:1 sparse:1 departing:1 feedback:1 world:1 avoids:1 contour:2 forward:8 made:1 adaptive:1 projected:3 lz:1 far:1 approximate:7 silhouette:4 ml:1 global:1 assumed:2 alternatively:1 factorizing:1 iterative:3 latent:1 promising:1 learn:4 kanade:1 improving:1 complex:3 artificially:1 domain:8 lvi:2 did:1 linearly:2 arrow:1 whole:1 body:26 eural:1 fig:5 referred:1 tl:1 fashion:1 sub:3 position:3 exponential:1 ib:1 specific:5 gating:1 normalizing:1 intractable:1 exists:2 workshop:1 illustrates:1 justifies:1 sclaroff:3 boston:4 led:1 photograph:1 simply:2 likely:5 visual:6 lagrange:1 tracking:10 truth:2 dh:1 ma:2 conditional:5 goal:1 formulated:1 careful:1 towards:2 hard:1 except:1 principal:1 called:2 brand:1 occluding:2 formally:2 people:2 incorporate:1 tested:1
1,118
202
490 Bell Learning in higher-order' artificial dendritic trees' Tony Bell Artificial Intelligence Laboratory Vrije Universiteit Brussel Pleinlaan 2, B-1050 Brussels, BELGIUM (tony@arti.vub.ac.be) ABSTRACT If neurons sum up their inputs in a non-linear way, as some simula- tions suggest, how is this distributed fine-grained non-linearity exploited during learning? How are all the small sigmoids in synapse, spine and dendritic tree lined up in the right areas of their respective input spaces? In this report, I show how an abstract atemporal highly nested tree structure with a quadratic transfer function associated with each branchpoint, can self organise using only a single global reinforcement scalar, to perform binary classification tasks. The procedure works well, solving the 6-multiplexer and a difficult phoneme classification task as well as back-propagation does, and faster. Furthermore, it does not calculate an error gradient, but uses a statistical scheme to build moving models of the reinforcement signal. 1. INTRODUCTION The computational territory between the linearly summing McCulloch-Pitts neuron and the non-linear differential equations of Hodgkin & Huxley is relatively sparsely populated. Connectionists use variants of the former and computational neuroscientists struggle with the exploding parameter spaces provided by the latter. However, evidence from biophysical simulations suggests that the voltage transfer properties of synapses, spines and dendritic membranes involve many detailed non-linear interactions, not just a squashing function at the cell body. Real neurons may indeed be higher-order nets. For the computationally-minded, higher order interactions means, first of all, quadratic terms. This contribution presents a simple learning principle for a binary tree with a logistic/quadratic transfer function at each node. These functions, though highly nested, are shown to be capable of changing their shape in concert. The resulting tree structure receives inputs at its leaves, and outputs an estimate of the probability that the input pattern is a member of one of two classes at the top. Learning in Higher-Order' Artificial Dendritic Trees' A number of other schemes exist for learning in higher-order neural nets. Sigma-Pi units, higher-order threshold logic units (Giles & Maxwell, 87) and product units (Durbin & Rumelhart, 89) are all examples of units which learn coefficients of non-linear functions. Product unit networks, like Radial Basis Function nets, consist of a layer of non-linear transformations, followed by a normal Perceptron-style layer. The scheme presented here has more in common with the work reviewed in Barron (88) (see also Tenorio 90) on polynomial networks in that it uses low order polynomials in a tree of low degree. The differences lie in a global rather than layer-by-Iayer learning scheme, and a transfer function derived from a gaussian discriminant function. 2. THE ARTIFICIAL DENDRITIC TREE (ADT) The network architecture in Figure I(a) is that of a binary tree which propagates real number values from its leaf nodes (or inputs) to its root node which is the output. In this simple formulation, the tree is construed as a binary classifier. The output node signals a number between 1 and 0 which represents the probability that the pattern presented to the tree was a member of the positive class of patterns or the negative class. Because the input patterns may have extremely high dimension and the tree is, at least initially, constrained to be binary, the depth of the tree may be significant, at least more than one might like to back-propagate through. A transfer function is associated with each 'hidden' node of the tree and the output node. This will hereafter be referred to as a Z{unction, for the simple reason that it takes in two variables X and Y, and outputs Z. A cascade of Z-functions performs the computation of the tree and the learning procedure consists of changing these functions. The tree is referred to as an Artificial Dendritic Tree or ADT with the same degree of licence that one may talk of Artificial Neural Networks, or ANNs. (a) z (x ,y) 1.0 z (x) (b) I A (c) x lnput nodes (d) X Y Figure 1: (a) an Artificial Dendritic Tree, (b) a ID Z-node (c) a 2D Z-node (d) A ID Z-function constructed from2 gaussians (e) approximating a step function 2.1. THE TRANSFER FUNCTION The idea behind the Z-function is to allow the two variables arriving at a node to interact locally in a non-linear way which contributes to the global computation of the tree. The transfer function is derived from statistical considerations. To simplify, consider the one-dimensional case of a variable X travelling on a wire as in Figure 1(b). A statistical estimation procedure could observe the distribution of values of X when the global pattern was positive or negative and derive a decision rule from these. In Figure I(d), the two density functions f+(x) and f-(x) are plotted. Where they meet, the local computation must answer that, based on its information, the global pattern is positively classified with probability 0.5. Assuming that there are equal numbers of positive and negative patterns (ie: that the a priori probability of positive is 0.5), it is easy to see that the conditional probability of being in the positive class given our value for X, is given by equation (1). 491 492 Bell z (x) = P [class=+ve Ix] = [+ex) [+(x)+[-(x) (1) This can be also derived from Bayesian reasoning (Therrien, 89). The fonn of z (x) is shown with the thick line in Figure l(d) for the given [+(x) and [-(x). If [+(x) and [-ex) can be usefully approximated by normal (gaussian) curves as plotted above, then (1) translates into (2): z ex) = 1. 1+e -mp" t ,input = ~-(x) - ~+(x) + In[ a:] a (2) This can be obtained by substituting equation (4) overleaf into (1) using the definitions of a and ~ given. The exact form a and ~ take depends on the number of variables input. The first striking thing is that the form of (2) is exactly that of the backpropagation logistic function. The second is that input is a polynomial quadratic expression. For Z-functions with 2 inputs (x ,y) using formulas (4.2) it takes the fonn: w lX2+W2Y2+w~+w 4X+wsY+w6 (3) The w' s can be thought of as weights just as in backprop, defining a 6D space of transfer functions. However optimising the w's directly through gradient descent may not be the best idea (though this is what Tenorio does), since for any error function E, law 4 = x law 1 =Y law 3. That is, the axes of the optimisation are not independent of each other. There are, however, two sets of 5 independent parameters which the w's in (3) are actually composed from if we calculate input from (4.2). These are Jl:, cr;, cr; and r+, denoting the means, standard deviations and correlation coefficient defining the two-dimensional distribution of (x ,y) values which should be positively classified. The other 5 variables define the negative distribution. Thus 2 Gaussians (hereafter referred to as the positive and negative models) define a quadratic transfer function (called the Z{unction) which can be interpreted as expressing conditional probability of positive class membership. The shape of these functions can be altered by changing the statistical parameters defining the distributions which undedy them. In Figure l(d), a 1-dimensional Z-function is seen to be sigmoidal though it need not be monotonic at all. Figure 2(b)-(h) shows a selection of 2D Zfunctions. In general the Z-function divides its N-dimensional input space with a N-1 dimensional hypersurface. In 2D, this will be an ellipse, a parabola, a hyperbola or some combination of the three. Although the dividing surface is quadratic, the Zfunction is still a logistic or squashing function. The exponent input is actually equivalent to the log likelihood ratio or In(j+(x)/j-(x?. commonly used in statistics. In this work, 2-dimensional gaussians are used to generate Z-functions. There are compelling reasons for this. One dimensional Z-functions are of little use since they do not reduce information. Z-functions of dimension higher than 1 perform optimal class-based information reduction by propagating conditional probabilities of class membership. But 2D Z-functions using 2D gaussians are of particular interest because they include in their function space all boolean functions of two variables (or at least analogue versions of these functions). For example the gaussians which would come to represent the positive and negative exemplar patterns for XOR are drawn as ellipses in Figure 2(a). They have equal means and variances but the negative exemplar patterns are correlated while the positive ones are anti-correlated. These models automatically give rise to the XOR surface in Figure 2(b) if put through equation (2). An interesting aE aE 11;, aE Learning in Higher-Order' Artificial Dendritic Trees' observation is that a problem of Nth order (XOR is 2nd order, 3-parity is 3rd order etc) can be solved by a polynomial of degree N (Figure 2d). Since 2nd degree polynomials like (3) are used in our system, there is one step up in power from 1st degree systems like the Perceptron. Thus 3-parity is to the Z-function unit what XOR is to the Perceptron (in this case not quadratically separable). A GAUSSIAN IS: f (x)=.le-IJ(%) a (4) in one dimension: a=(21t) 1120'% (4.1.1) ~(x ) (4.1.2) (x -Jl% )2 20'x 2 in two dimensions: a=21tO'x O'y(l-r 2)112 1 ~(x ,y)= 2(l-r2) in n dimensions: (4.2.1) [ (X-J,1x)2 0'% 2 + (y -~ )2 2r (x -J,1x )(y -~ ) 0'/ O'x O'y a=(21t)" /2 IK 11/2 (4.n.l) ~<!)= ~ (!-mlK- 1<!-m) (4.n.2) Jl%=E [x] is the expected value or mean of x O';:E [x 2]-Jl% 2 is the variance of x r E[xy]~%Jly 1 (4.2.2) is the correlation coefficiem of a bivariate gaussian m=E [!] is the mean vector of a multivariate gaussian K=E [<!-m)<!-m)T] is the covariance matrix of a multivariate? gaussian with IK 1its determinant (j) (i) (k) /l\. M Figure 2: (a) two anti-(;orrelated gaussians seen from above (b) the resulting Zfunction (c)-(h) Some other 20 Z-functions. (i) 3-parity in a cube cannot be solved by a 30 Z-function (j) but yields to a cascade of 20 ones (k). 2.2. THE LEARNING PROCEDURE If gaussians are used to model the distribution of inputs x which give positive and negative classification errors, rather than just the distribution of positively and negatively classified x, then it is possible to formulate an incremental learning procedure for training Z-functions. This procedure enables the system to deal with data which is not gaussianly distributed. 493 494 Bell 2.2.1. Without hidden units: learning a step function. A simple example illustrates this principle. Consider a network consisting entirely of a I-dimensional Z-function. as in Figure 1(b). The input is a real number from 0 to 1 and the output is to be a step function, such that 0.5-1.0 is classed positively (output 1.0) and 0.0-0.5 should output 0.0. The 4 parameters of the Z-function (Jl+,Jl-,cr+,crl are initialised randomly and example patterns are presented to the 'tree'. On each presentation t, the error 0 in the response is calculated by 0, ~ d, -0" the desired minus the actual output at time t, and 2 of the parameters are altered. If the error is positive, the positive model is altered, otherwise the negative model is altered. Changing a model consists of 'sliding' the estimates of the appropriate first and second moments (E[x] and E [x 2]) according to a 'moving-average' scheme: E [x], ~ to,x,+(1-to,)E [x ]'-1 (5.1) (5.2) where t is a plasticity or learning rate, x, is the value input and E [x ]'-1 was the previous estimate of the mean value of x for the appropriate gaussian. This rule means that at any moment, the parameters determining the positive and negative models are weighted averages of recent inputs which have generated errors. The influence which a particular input has had decays over time. This algorithm was run with ?=0.1. After 100 random numbers had been presented, with error signals from the step-function changing the models, the models come to well represent the distribution of positive and negative inputs. At this stage the models and their associated Z-function are those shown in Figure l(d). But now, most of the error reinforcement will be coming from a small region around 0.5, which means that since the gaussians are modelling the errors, they will be drawn towards the centre and become narrower. This has the effect, Figure l(e), of increasing the gain of the sigmoidal Z-function. In the limit, it will converge to a perfect step function as the gaussians become infinitesimally separated delta functions. This initial demonstration shows the automatic gain adjustment property of the Z-function. 2.2.2. With hidden units: the 6-multiplexer. The first example showed how a 1D Z-function can minimise error by modelling it. This example shows how a cascade of 2-dimensional Z-functions can co-operate to solve a 3rd order problem. A 6-multiplexer circuit receives as input 6 bits, 4 of which are data bits and 2 are address bits. If the address bits are 00, it must output the contents of the first data bit, if 01, the second, 10 the third and 11 the fourth. There are 64 different input patterns. Choosing an tree architecture is a difficult problem in general, but the first step is to choose one which we know can solve the problem. This is illustrated in Figure 3(a). This is an architecture for which there exists a solution using binary Boolean functions. The tree's solution was arrived at as follows: each node was initialised with 10 random values: E[x]' E[y], E[x 2], E[y2] and E[xy] for each of its positive and negative models. The learning rate t was set to 0.02 and input patterns were generated and propagated up to the top node, where an error measurement was made. The error was then broadcast globally to all nodes, each one, in effect, being told to respond more positively (or negatively) should the same circumstances arise again, and adjusting their Z-functions in the same way as equations (5). This time, however, 5 parameters Learning in Higher-Order' Artificial Dendritic Trees' were adjusted per node per presentation. instead of 2. Again. which model (positive or negative) is adjusted depends on the sign of the error at the top of the tree. The tree learns after about 200 random bit patterns are presented (7 seconds on a Symbolics). After 300 presentations (the state depicted in Figure 3a), the mean squared error is falling steadily to zero. An adequate back-propagation network takes 6000 presentations to converge on a solution. The solution achieved is a rather messy combination of half-hearted XORs and NXORs, and ambiguous AND/ORs. The problem was tried with different trees. In general any tree of sufficient richness can solve the problem though larger trees take longer. Trees for which no nice solutions exist. ie: those with fewer than 6 well-chosen inputs from the address bits can sometimes still perform rather well. A tree with straight convergence. only one contact per address bit, can still quickly approach 80% performance, but further training is destructive. Figure 3(b) shows a tree trained to output 1 if half or more of its 8 inputs were on. Al rr===---n 7 ... (a) 8 "~_--'I (b) Figure 3: Solving the 6-multiplexer (a) and the 8-majority predicate (b) 2.2.3. Phoneme classification. A good question was if such a tree could perform well on a large problem, so a typical back-propagation application was attempted. Space does not permit a full account here. but the details appear in Bell (89). The data came from 100 speakers speaking the confusable E-set phonemes (B, D, E and V). This was the same data as that used by Lang & Hinton (88). Four trees were built out of 192 input units and the trees trained using a learning schedule of E falling from 0.01 to 0.001 over the course of 30 presentations of each of 668 training patterns. Generalisation to a test set was 88.5%, 0.5% worse than an equivalently simple backprop net A more sophisticated backprop net, using time-delays and multiresolution training could reach 93% generalisation. Thirty epochs with the trees took some 16 hours on a Sun 3-260 whereas the backprop experiments were performed on a Convex supercomputer. The conclusion from these experiments is that trees some 8 levels deep are capable of almost matching normal back-propagation on a large classification task in a fraction of the training time. Attempts to build time-symmetry into the trees have not so far been successful. 3. DISCUSSION Even within the context of other connectionist leaming procedures, there is something of an air of mystery about this one. The apparatus of gradient descent, either for individual units or for the whole tree is absent or at least hidden. 495 496 Bell 3.1. HOW DOES IT WORK? It is necessary to reflect on the effect of modelling errors. Models of errors are an attempt to push a node's outputs towards the edge of its parent's input square. Where the model is perfect, it is simple for the node above to model the model by applying a sigmoid, and so on to the top of the tree, where the error disappears. But the modelling is actually done in a totally distributed and collaborative way. The identification of 1.0 with positive error (top output too small) means that Z-functions are more likely to be monotonic towards (1,1) the further they are from the inputs. Two standard problems are overcome in unusual ways. The first, credit assignment, is solved because different Z-functions are able to model different errors, giving them different roles. Although all nodes receive the same feedback, some changes to a node's model will be swiftly undone when the new errors that result from them begin to be broadcast. Other nodes can change freely either because they are not yet essential to the computation or because there exist alterations of their models tolerable to the nodes above. The second problem is stability. In backprop, the way the error diffuses through the net ensures that the upper weights are slaved to the lower ones because the lower are changing more slowly. In this system, the upper nodes are slaved to the lower ones because they are explicitly modelling their activities. Conversely, the lower nodes will never be allowed to change too quickly since the errors generated by sluggish top nodes will throw them back into the behaviour the top nodes expect For a low enough learning rate e, the solutions are stable. Amongst the real problems with this system are the following. First, the credit assignment is not solved for units receiving the same input variables, making many normal connectionist architectures impossible. Second, the system can only deal with 2 classes. Third, as with other algorithms, choice of architecture is a 'black art'. 3.2. BIOPHYSICS & REAL NEURONS The name' Artificial Dendritic Tree' is perhaps overdoing it. The tree has no dynamic properties, activation flows in only one direction, the branchpoints of the tree routinely implement XOR and the 'cell' as a whole implements phoneme recognition (only a small step from grandmothers). The title was kept because what drove the work was a search for a computational explanation of how fine-grained local non-linearities of low degree could combine in a learning process. Work in computational neuroscience, in particular with compartmental models (Koch & Poggio 87; RaIl & Segev 88; Segev et al 89, Shepherd & Brayton 87) have shown that it is likely that many non-linear effects take place between synapse and soma. Synaptic transfer functions can be sigmoidal, spines with active channels may mutually excite each other (even implement boolean computations) and inhibitory inputs can 'veto' firing in a highly non-linear fashion (silent inhibition). The dendritic membrane itself is filled with active ion channels, whose boosting or quenching properties depend in a complex way on the intracellular voltage levels or Ca'Jn. concentration (itself dependent on voltage). Thus we may be able to consider the membrane itself as a distributed processing system, meaning that the synapses are no longer the privileged sites of learning which they have tended to be since Hebb. Active channels can serve to implement threshold functions just as well at the dendritic branchpoints as at the soma, where they generate spikes. There are many different kinds of ion channel (Yamada et aI, 89) with inhomogenous distributions over the dendritic tree. A neuron's DNA may generate a certain 'base set' of channel proteins that span a non-linear function space just as our Learning in Higher-Order' Arti ficial Dendritic Trees' parameters span the Z-function space. The properties of a part of dendritic membrane could be seen as a point in channel space. Viewed this way. the neuron becomes one large computer. When one considers the Purkinje cell of the cerebellum with 100.000 inputs, as many spines. a massive arborisation full of active channels, many of them Ca-permeable or Ca-dependent. with spiking and plateau potentials occurring in the dendritic tree. the notion that the cell may be implementing a 99.999 dimensional hyperplane starts to recede. here is an extra motivation for considering the cell as a complex computer. Algorithms such as back-propagation would require feedback circuits to send error. If the cell is the feedback unit, then reinforcement can occur as a spike at the soma rein vades the dendritic tree. Thus nerves may not spike just for axonal purposes. but also to penetrate the electrotonic length of the dendrites. This was thought to be a component of Hebbian learning at the synapses, but it could be the basis of more if the dendritic membrane computes. 4. Acknowledgements To Kevin Lang for the speech data and to Rolf Pfeifer and Luc Steels for support. Further credits in Bell (90). The author is funded by ESPRIT B.R.A. 3234. 5. References Barron A & Barron R (88) Statistical Learning Networks: a unifying view, in Wegman E (ed) Proc. 20th Symp. on Compo Science & Statistics [see also this volume] Bell T (89) Artificial Dendritic Learning. in Almeida L. (ed) Proc. EURASIP Workshop on Neural Networks. Lecture notes in Computer Science. SpringerVerlag. [also VUB AI-lab Memo 89-20]. Durbin R & Rumelhart D (89) Product Units: A Computationally Powerful and Biologically Plausible Extansion to Backpropagation Nets. Neural Computation J Giles C.L. & Maxwell T (87) Learning. in variance and generalisation in high-order neural networks. Applied Optics vol 26. no. 23 Koch C & Poggio T (87) Biophysics of Computational Systems: Neurons, synapses and membranes. in G. Edelman et al (eds). Synaptic Function. John Wiley. Lang K & Hinton G (88) The Development of the Time-Delay Neural Network Architecture for Speech Recognition. Tech Report CMU-CS-88-J52 RaIl W & Segev I (88) Excitable Dendritic Spine Clusters: non-linear synaptic processing. in R.Cotterill (ed) Computer Simulation in Brain Science. Camb.U.P. Segev I. Fleshman J & Burke R. (89) Compartmental Models of Complex Neurons. in Methods in Neuronal Modelling Shepherd G & Brayton R (87) Logic operations are properties of computer simulated interactions between excitable dendritic spines. Neuroscience. vol 21, no. 1 1987 Koch C & Segev I (eds) MIT press 1989 Tenorio M & Lee W (90) Self-Organizing Network for Optimal Supervised Learning, IEEE Transactions in Neural Networks, 1990 [see also this volume] Therrien C (89) Decision Estimation and Classification. Yamada W, Koch C & Adams P (89) Multiple Channels and Calcium Dynamics, in Methods in Neuronal Modelling Koch C & Segev I (eds) MIT press 1989. 497
202 |@word determinant:1 version:1 polynomial:5 nd:2 simulation:2 propagate:1 tried:1 covariance:1 arti:2 fonn:2 minus:1 mlk:1 moment:2 reduction:1 initial:1 hereafter:2 denoting:1 unction:2 lang:3 yet:1 activation:1 must:2 john:1 plasticity:1 shape:2 enables:1 concert:1 intelligence:1 leaf:2 half:2 fewer:1 ficial:1 yamada:2 compo:1 boosting:1 node:24 sigmoidal:3 constructed:1 differential:1 become:2 ik:2 edelman:1 consists:2 lx2:1 combine:1 symp:1 jly:1 expected:1 indeed:1 spine:6 brain:1 globally:1 automatically:1 little:1 actual:1 considering:1 increasing:1 totally:1 provided:1 begin:1 linearity:2 becomes:1 circuit:2 mcculloch:1 what:3 kind:1 interpreted:1 transformation:1 usefully:1 exactly:1 esprit:1 classifier:1 unit:13 appear:1 positive:17 local:2 struggle:1 limit:1 apparatus:1 id:2 permeable:1 meet:1 firing:1 might:1 black:1 suggests:1 conversely:1 co:1 thirty:1 implement:4 backpropagation:2 procedure:7 area:1 bell:8 cascade:3 thought:2 matching:1 undone:1 radial:1 suggest:1 protein:1 cannot:1 selection:1 put:1 context:1 influence:1 applying:1 impossible:1 equivalent:1 send:1 convex:1 formulate:1 penetrate:1 rule:2 stability:1 notion:1 drove:1 massive:1 exact:1 us:2 simula:1 rumelhart:2 approximated:1 recognition:2 parabola:1 sparsely:1 role:1 solved:4 calculate:2 region:1 ensures:1 richness:1 sun:1 messy:1 dynamic:2 trained:2 depend:1 solving:2 serve:1 negatively:2 basis:2 routinely:1 talk:1 separated:1 artificial:11 kevin:1 choosing:1 rein:1 adt:2 whose:1 larger:1 solve:3 plausible:1 compartmental:2 otherwise:1 statistic:2 itself:3 rr:1 biophysical:1 net:7 took:1 interaction:3 product:3 coming:1 lnput:1 organizing:1 multiresolution:1 convergence:1 parent:1 cluster:1 incremental:1 perfect:2 arborisation:1 adam:1 tions:1 derive:1 ac:1 propagating:1 exemplar:2 ij:1 ex:3 dividing:1 throw:1 brayton:2 c:1 come:2 direction:1 thick:1 implementing:1 backprop:5 require:1 behaviour:1 licence:1 dendritic:21 adjusted:2 burke:1 pleinlaan:1 around:1 credit:3 normal:4 koch:5 pitt:1 substituting:1 quenching:1 belgium:1 purpose:1 estimation:2 proc:2 title:1 minded:1 weighted:1 mit:2 gaussian:7 rather:4 cr:3 voltage:3 derived:3 ax:1 modelling:7 likelihood:1 tech:1 dependent:2 membership:2 initially:1 hidden:4 diffuses:1 classification:6 priori:1 exponent:1 development:1 constrained:1 art:1 cube:1 equal:2 never:1 optimising:1 represents:1 report:2 connectionist:2 simplify:1 randomly:1 composed:1 ve:1 individual:1 vrije:1 consisting:1 attempt:2 neuroscientist:1 interest:1 highly:3 cotterill:1 behind:1 edge:1 capable:2 necessary:1 xy:2 poggio:2 respective:1 tree:47 filled:1 divide:1 desired:1 plotted:2 confusable:1 vub:2 giles:2 compelling:1 boolean:3 purkinje:1 assignment:2 deviation:1 delay:2 predicate:1 successful:1 too:2 answer:1 swiftly:1 st:1 density:1 ie:2 therrien:2 told:1 receiving:1 lee:1 quickly:2 again:2 squared:1 reflect:1 choose:1 broadcast:2 slowly:1 worse:1 style:1 multiplexer:4 account:1 potential:1 alteration:1 coefficient:2 mp:1 explicitly:1 depends:2 performed:1 root:1 view:1 lab:1 universiteit:1 start:1 contribution:1 construed:1 square:1 air:1 collaborative:1 xor:5 phoneme:4 variance:3 yield:1 hyperbola:1 territory:1 bayesian:1 identification:1 straight:1 classified:3 anns:1 plateau:1 synapsis:4 reach:1 tended:1 ed:6 synaptic:3 definition:1 initialised:2 steadily:1 destructive:1 associated:3 propagated:1 gain:2 adjusting:1 schedule:1 sophisticated:1 actually:3 back:7 nerve:1 maxwell:2 higher:10 supervised:1 response:1 synapse:2 formulation:1 done:1 though:4 furthermore:1 just:6 stage:1 correlation:2 receives:2 propagation:5 logistic:3 perhaps:1 name:1 effect:4 y2:1 former:1 laboratory:1 illustrated:1 deal:2 cerebellum:1 during:1 self:2 ambiguous:1 speaker:1 arrived:1 performs:1 reasoning:1 meaning:1 consideration:1 common:1 sigmoid:1 spiking:1 camb:1 volume:2 jl:6 significant:1 expressing:1 measurement:1 ai:2 rd:2 automatic:1 populated:1 centre:1 had:2 funded:1 moving:2 stable:1 longer:2 surface:2 inhibition:1 etc:1 base:1 something:1 multivariate:2 recent:1 showed:1 certain:1 binary:6 came:1 exploited:1 seen:3 freely:1 converge:2 signal:3 exploding:1 sliding:1 full:2 multiple:1 hebbian:1 faster:1 ellipsis:1 biophysics:2 privileged:1 variant:1 ae:3 optimisation:1 circumstance:1 cmu:1 represent:2 sometimes:1 achieved:1 cell:6 ion:2 receive:1 whereas:1 fine:2 extra:1 operate:1 shepherd:2 thing:1 member:2 veto:1 flow:1 axonal:1 easy:1 enough:1 architecture:6 silent:1 reduce:1 idea:2 translates:1 minimise:1 absent:1 expression:1 speech:2 speaking:1 adequate:1 deep:1 electrotonic:1 detailed:1 involve:1 brussel:1 locally:1 dna:1 generate:3 exist:3 inhibitory:1 sign:1 delta:1 neuroscience:2 per:3 vol:2 four:1 soma:3 threshold:2 falling:2 drawn:2 changing:6 kept:1 fraction:1 sum:1 run:1 mystery:1 fourth:1 respond:1 hodgkin:1 striking:1 powerful:1 place:1 almost:1 decision:2 bit:8 entirely:1 layer:3 followed:1 quadratic:6 durbin:2 activity:1 occur:1 optic:1 huxley:1 segev:6 extremely:1 span:2 separable:1 infinitesimally:1 relatively:1 according:1 brussels:1 combination:2 membrane:6 making:1 biologically:1 inhomogenous:1 computationally:2 equation:5 mutually:1 lined:1 know:1 unusual:1 travelling:1 gaussians:9 operation:1 permit:1 observe:1 barron:3 appropriate:2 tolerable:1 supercomputer:1 jn:1 top:7 symbolics:1 tony:2 include:1 unifying:1 giving:1 build:2 ellipse:1 approximating:1 contact:1 question:1 spike:3 concentration:1 gradient:3 amongst:1 simulated:1 majority:1 considers:1 discriminant:1 reason:2 assuming:1 w6:1 length:1 ratio:1 demonstration:1 equivalently:1 difficult:2 classed:1 sigma:1 negative:13 rise:1 steel:1 memo:1 calcium:1 xors:1 perform:4 upper:2 neuron:8 wire:1 observation:1 descent:2 anti:2 wegman:1 defining:3 hinton:2 connectionists:1 quadratically:1 hour:1 address:4 able:2 pattern:14 grandmother:1 rolf:1 built:1 explanation:1 analogue:1 power:1 nth:1 scheme:5 altered:4 disappears:1 excitable:2 nice:1 epoch:1 acknowledgement:1 determining:1 law:3 expect:1 lecture:1 interesting:1 degree:6 sufficient:1 propagates:1 principle:2 pi:1 squashing:2 course:1 parity:3 arriving:1 organise:1 allow:1 perceptron:3 distributed:4 curve:1 dimension:5 depth:1 calculated:1 overcome:1 feedback:3 computes:1 author:1 commonly:1 reinforcement:4 made:1 far:1 hypersurface:1 transaction:1 logic:2 global:5 active:4 summing:1 excite:1 iayer:1 search:1 reviewed:1 channel:8 learn:1 transfer:10 ca:3 symmetry:1 contributes:1 dendrite:1 interact:1 complex:3 linearly:1 intracellular:1 whole:2 motivation:1 arise:1 allowed:1 body:1 positively:5 site:1 referred:3 neuronal:2 fashion:1 hebb:1 wiley:1 lie:1 rail:2 third:2 pfeifer:1 ix:1 grained:2 learns:1 formula:1 r2:1 decay:1 evidence:1 bivariate:1 consist:1 exists:1 essential:1 workshop:1 sluggish:1 sigmoids:1 illustrates:1 push:1 occurring:1 depicted:1 likely:2 tenorio:3 adjustment:1 scalar:1 monotonic:2 slaved:2 nested:2 conditional:3 viewed:1 presentation:5 narrower:1 towards:3 leaming:1 luc:1 crl:1 content:1 change:3 eurasip:1 springerverlag:1 typical:1 generalisation:3 hyperplane:1 called:1 attempted:1 support:1 almeida:1 latter:1 correlated:2
1,119
2,020
On Discriminative vs. Generative classifiers: A comparison of logistic regression and naive Bayes Andrew Y. Ng Michael I. Jordan Computer Science Division C.S. Div. & Dept. of Stat. University of California, Berkeley University of California, Berkeley Berkeley, CA 94720 Berkeley, CA 94720 Abstract We compare discriminative and generative learning as typified by logistic regression and naive Bayes. We show, contrary to a widelyheld belief that discriminative classifiers are almost always to be preferred, that there can often be two distinct regimes of performance as the training set size is increased, one in which each algorithm does better. This stems from the observation- which is borne out in repeated experiments- that while discriminative learning has lower asymptotic error, a generative classifier may also approach its (higher) asymptotic error much faster. 1 Introduction Generative classifiers learn a model of the joint probability, p( x, y), of the inputs x and the label y, and make their predictions by using Bayes rules to calculate p(ylx), and then picking the most likely label y. Discriminative classifiers model the posterior p(ylx) directly, or learn a direct map from inputs x to the class labels. There are several compelling reasons for using discriminative rather than generative classifiers, one of which, succinctly articulated by Vapnik [6], is that "one should solve the [classification] problem directly and never solve a more general problem as an intermediate step [such as modeling p(xly)]." Indeed, leaving aside computational issues and matters such as handling missing data, the prevailing consensus seems to be that discriminative classifiers are almost always to be preferred to generative ones. Another piece of prevailing folk wisdom is that the number of examples needed to fit a model is often roughly linear in the number of free parameters of a model. This has its theoretical basis in the observation that for "many" models, the VC dimension is roughly linear or at most some low-order polynomial in the number of parameters (see, e.g., [1, 3]), and it is known that sample complexity in the discriminative setting is linear in the VC dimension [6]. In this paper, we study empirically and theoretically the extent to which these beliefs are true. A parametric family of probabilistic models p(x, y) can be fit either to optimize the joint likelihood of the inputs and the labels, or fit to optimize the conditional likelihood p(ylx), or even fit to minimize the 0-1 training error obtained by thresholding p(ylx) to make predictions. Given a classifier hGen fit according to the first criterion, and a model h Dis fit according to either the second or the third criterion (using the same parametric family of models) , we call hGen and h Dis a Generative-Discriminative pair. For example, if p(xly) is Gaussian and p(y) is multinomial, then the corresponding Generative-Discriminative pair is Normal Discriminant Analysis and logistic regression. Similarly, for the case of discrete inputs it is also well known that the naive Bayes classifier and logistic regression form a Generative-Discriminative pair [4, 5]. To compare generative and discriminative learning, it seems natural to focus on such pairs. In this paper, we consider the naive Bayes model (for both discrete and continuous inputs) and its discriminative analog, logistic regression/linear classification, and show: (a) The generative model does indeed have a higher asymptotic error (as the number of training examples becomes large) than the discriminative model, but (b) The generative model may also approach its asymptotic error much faster than the discriminative model- possibly with a number of training examples that is only logarithmic, rather than linear, in the number of parameters. This suggests-and our empirical results strongly support-that, as the number of training examples is increased, there can be two distinct regimes of performance, the first in which the generative model has already approached its asymptotic error and is thus doing better, and the second in which the discriminative model approaches its lower asymptotic error and does better. 2 Preliminaries We consider a binary classification task, and begin with the case of discrete data. Let X = {O, l}n be the n-dimensional input space, where we have assumed binary inputs for simplicity (the generalization offering no difficulties). Let the output labels be Y = {T, F}, and let there be a joint distribution V over X x Y from which a training set S = {x(i) , y(i) }~1 of m iid examples is drawn. The generative naive Bayes classifier uses S to calculate estimates p(xiIY) and p(y) of the probabilities p(xi IY) and p(y), as follows: ,y=b}+1 (1) P' (x-, = 11Y = b) = #s{xi=l #s{y-b}+21 (and similarly for p(y = b),) where #s{-} counts the number of occurrences of an event in the training set S. Here, setting l = corresponds to taking the empirical estimates of the probabilities, and l is more traditionally set to a positive value such as 1, which corresponds to using Laplace smoothing of the probabilities. To classify a test example x, the naive Bayes classifier hGen : X r-+ Y predicts hGen(x) = T if and only if the following quantity is positive: ? IGen(x ) = log (rr~-d) (x i ly = T))p(y = T) (rrni=1 P' (X,_IY -_ F)) P'( Y -_ ~ p(xilY = T) F) = L..,log ' ( _I _ F) i=1 P X, Y - p(y = T) + log P' (Y -_ F)' (2) In the case of continuous inputs , almost everything remains the same, except that we now assume X = [O,l]n, and let p(xilY = b) be parameterized as a univariate Gaussian distribution with parameters {ti ly=b and if; (note that the j1's, but not the if's , depend on y). The parameters are fit via maximum likelihood, so for example {ti ly=b is the empirical mean of the i-th coordinate of all the examples in the training set with label y = b. Note that this method is also equivalent to Normal Discriminant Analysis assuming diagonal covariance matrices. In the sequel, we also let J.tily=b = E[Xi IY = b] and = Ey[Var(xi ly)] be the "true" means and variances (regardless of whether the data are Gaussian or not). In both the discrete and the continuous cases, it is well known that the discriminative analog of naive Bayes is logistic regression. This model has parameters [,8, OJ, and posits that p(y = Tlx; ,8, O) = 1/(1 +exp(-,8Tx - 0)). Given a test example x, a; the discriminative logistic regression classifier ho is : X and only if the linear discriminant function I-t Y predicts hOis (x) = T if lDis(x) = L~=l (3ixi + () (3) is positive. Being a discriminative model, the parameters [(3, ()] can be fit either to maximize the conditionallikelikood on the training set L~= llogp(y(i) Ix(i); (3, ()), or to minimize 0-1 training error L~= ll{hois(x(i)) 1- y(i)}, where 1{-} is the indicator function (I{True} = 1, I{False} = 0) . Insofar as the error metric is 0-1 classification error, we view the latter alternative as being more truly in the "spirit" of discriminative learning, though the former is also frequently used as a computationally efficient approximation to the latter. In this paper, we will largely ignore the difference between these two versions of discriminative learning and, with some abuse of terminology, will loosely use the term "logistic regression" to refer to either, though our formal analyses will focus on the latter method. Finally, let 1i be the family of all linear classifiers (maps from X to Y); and given a classifier h : X I-t y, define its generalization error to be c(h) = Pr(x,y)~v [h(x) 1- y]. 3 Analysis of algorithms When V is such that the two classes are far from linearly separable, neither logistic regression nor naive Bayes can possibly do well, since both are linear classifiers. Thus, to obtain non-trivial results, it is most interesting to compare the performance of these algorithms to their asymptotic errors (cf. the agnostic learning setting). More precisely, let hGen,oo be the population version of the naive Bayes classifier; i.e. hGen,oo is the naive Bayes classifier with parameters p(xly) = p(xly),p(y) = p(y). Similarly, let hOis ,oo be the population version of logistic regression. The following two propositions are then completely straightforward. Proposition 1 Let hGen and h Dis be any generative-discriminative pair of classifiers, and hGen,oo and hois, oo be their asymptotic/population versions. Then l c(hDis,oo) :S c(hGen,oo). Proposition 2 Let h Dis be logistic regression in n-dimensions. Then with high probability c(hois ) :S c(hois,oo) + 0 (J ~ log ~) Thus, for c(hOis ) :S c(hOis,oo) + EO to hold with high probability (here, fixed constant), it suffices to pick m = O(n). EO > 0 is some Proposition 1 states that aymptotically, the error of the discriminative logistic regression is smaller than that of the generative naive Bayes. This is easily shown by observing that, since c(hDis) converges to infhE1-l c(h) (where 1i is the class of all linear classifiers), it must therefore be asymptotically no worse than the linear classifier picked by naive Bayes. This proposition also provides a basis for what seems to be the widely held belief that discriminative classifiers are better than generative ones. Proposition 2 is another standard result, and is a straightforward application of Vapnik's uniform convergence bounds to logistic regression, and using the fact that 1i has VC dimension n. The second part of the proposition states that the sample complexity of discriminative learning- that is, the number of examples needed to approach the asymptotic error- is at most on the order of n. Note that the worst case sample complexity is also lower-bounded by order n [6]. lUnder a technical assumption (that is true for most classifiers, including logistic regression) that the family of possible classifiers hOis (in the case of logistic regression, this is 1l) has finite VC dimension. The picture for discriminative learning is thus fairly well-understood: The error converges to that of the best linear classifier, and convergence occurs after on the order of n examples. How about generative learning, specifically the case of the naive Bayes classifier? We begin with the following lemma. ? ? Lemma 3 Let any 101,8 > and any l 2: be fixed. Assume that for some fixed Po > 0, we have that Po :s: p(y = T) :s: 1 - Po. Let m = 0 ((l/Ei) log(n/8)). Then with probability at least 1 - 8: 1. In case of discrete inputs, IjJ(XiIY = b) - p(xilY b) - p(y = b) I :s: 101, for all i = 1, ... ,n and bEY. = b)1 :s: 101 and IjJ(y = 2. In the case of continuous inputs, IPi ly=b - f-li ly=b I :s: 101, laT IjJ(y = b) - p(y = b) I :s: 101 for all i = 1, ... ,n and bEY. - O"T I :s: 101, and ? Proof (sketch). Consider the discrete case, and let l = for now. Let 101 :s: po/2. By the Chernoff bound, with probability at least 1 - 81 = 1- 2exp(-2Eim) , the fraction of positive examples will be within 101 of p(y = T) , which implies IjJ(y = b) - p(y = b)1 :s: 101, and we have at least 1 m positive and 1m negative examples, where I = Po - 101 = 0(1). So by the Chernoff bound again , for specific i, b, the chance that IjJ(XiIY = b) - p(xilY = b)1 > 101 is at most 82 = 2exp(-2Ehm). Since there are 2n such probabilities, the overall chance of error, by the Union bound, is at most 81 + 2n82 . Substituting in 81 and 8/s definitions , we see that to guarantee 81 + 2n82 :s: 8, it suffices that m is as stated. Lastly, smoothing (l > 0) adds at most a small, O(l/m) perturbation to these probabilities , and using the same argument as above with (say) 101/2 instead of 101, and arguing that this O(l/m) perturbation is at most 101/2 (which it is as m is at least order l/Ei) , again gives the result. The result for the continuous case is proved similarly using a Chernoff-bounds based argument (and the assumption that Xi E [0,1]). D Thus, with a number of samples that is only logarithmic, rather than linear, in n, the parameters of the generative classifier hGen are uniformly close to their asymptotic values in hGen ,oo . Is is tempting to conclude therefore that c(hGen), the error of the generative naive Bayes classifier, also converges to its asymptotic value of c(hGen,oo) after this many examples, implying only 0 (log n) examples are required to fit a naive Bayes model. We will shortly establish some simple conditions under which this intuition is indeed correct. Note that this implies that, even though naive Bayes converges to a higher asymptotic error of c(hGen,oo) compared to logistic regression's c: (hDis, oo ), it may also approach it significantly faster-after O(log n), rather than O(n), training examples. One way of showing c(hGen) approaches c(hGen,oo) is by showing that the parameters' convergence implies that hGen is very likely to make the same predictions as hGen,oo . Recall hGen makes its predictions by thresholding the discriminant function lGen defined in (2). Let lGen,oo be the corresponding discriminant function used by hGen,oo. On every example on which both lGen and lGen ,oo fall on the same side of zero, hGen and hGen,oo will make the same prediction. Moreover, as long as lGen,oo (x) is, with fairly high probability, far from zero, then lGen (x), being a small perturbation of lGen ,oo(x), will also be usually on the same side ofzero as lGen ,oo (x). Theorem 4 Define G(T) = Pr(x,y)~v[(lGen ,oo(x) E [O,Tn] A y = T) V (lG en,oo(X) E [-Tn, O]A Y = F)]. Assume that for some fixed Po > 0, we have Po :s: p(y = T) :s: 1 - Po, and that either Po :s: P(Xi = 11Y = b) :s: 1 - Po for all i, b (in the case of discrete inputs), or O"T 2: Po (in the continuous case). Then with high probability, c:( hGen ) :s: c:(hGen,oo) + G (0 (J ~ logn)) . (4) Proof (sketch). c(hGen) - c(hGen,oo) is upperbounded by the chance that hGen,oo correctly classifies a randomly chosen example, but hGen misclassifies it. Lemma 3 ensures that, with high probability, all the parameters of hGen are within O( j(log n)/m) of those of hGen ,oo . This in turn implies that everyone of the n + 1 terms in the sum in lGen (as in Equation 2) is within O( j(1ogn)/m) of the corresponding term in lGen ,oo , and hence that IlGen(x) -lGen,oo(x)1 :S O(nj(1ogn)/m). Letting T = O( j(logn)/m), we therefore see that it is possible for hGen,oo to be correct and hGen to be wrong on an example (x , y) only if y = T and lGen,oo(X) E [0, Tn] (so that it is possible that lGen,oo(X) ::::: 0, lGen (x) :S 0), or if y = F and lGen,oo(X) E [-Tn, 0]. The probability of this is exactly G(T), which therefore upperbounds c(hGen) - c(hGen,oo ). D The key quantity in the Theorem is the G(T) , which must be small when T is small in order for the bound to be non-trivial. Note G(T) is upper-bounded by Prx[lGen,oo(x) E [-Tn, Tn]]-the chance that lGen, oo(X) (a random variable whose distribution is induced by x ""' V) falls near zero. To gain intuition about the scaling of these random variables, consider the following: Proposition 5 Suppose that, for at least an 0(1) fraction of the features i (i = 1, ... ,n), it holds true that IP(Xi = 11Y = T) - P(Xi = 11Y = F)I ::::: 'Y for some fixed'Y > 0 (or IJLi ly=T - JLi ly=FI ::::: 'Y in the case of continuous inputs). Then E[lGen ,oo(x)ly = T] = O(n), and -E[lGen,oo (x)ly = F] = O(n). Thus, as long as the class label gives information about an 0(1) fraction of the features (or less formally, as long as most of the features are "relevant" to the class label), the expected value of IlGen, oo(X) I will be O(n). The proposition is easily proved by showing that, conditioned on (say) the event y = T, each of the terms in the summation in lGen, oo(x) (as in Equation (2), but with fi's replaced by p's) has non-negative expectation (by non-negativity of KL-divergence), and moreover an 0(1) fraction of them have expectation bounded away from zero. Proposition 5 guarantees that IlGen,oo (x)1 has large expectation, though what we want in order to bound G is actually slightly stronger, namely that the random variable IlGen,oo (x)1 further be large/far from zero with high probability. There are several ways of deriving sufficient conditions for ensuring that G is small. One way of obtaining a loose bound is via the Chebyshev inequality. For the rest of this discussion, let us for simplicity implicitly condition on the event that a test example x has label T. The Chebyshev inequality implies that Pr[lGen ,oo(x) :S E[lGen ,oo(X)] - t] :S Var(lGen,oo(x))/t2 . Now, lGen,oo (X) is the sum of n random variables (ignoring the term involving the priors p(y)). If (still conditioned on y), these n random variables are independent (i.e. if the "naive Bayes assumption," that the xi's are conditionally independent given y, holds), then its variance is O(n); even if the n random variables were not completely independent, the variance may still be not much larger than 0 (n) (and may even be smaller, depending on the signs of the correlations), and is at most O(n 2). So, if E[lGen,oo (x)ly = T] = an (as would be guaranteed by Proposition 5) for some a > 0, by setting t = (a - T)n, Chebyshev's inequality gives Pr[lGen,oo(x) :S Tn] :S O(l/(a - T)2n1/) (T < a), where 1} = 0 in the worst case, and 1} = 1 in the independent case. This thus gives a bound for G(T), but note that it will frequently be very loose. Indeed, in the unrealistic case in which the naive Bayes assumption really holds , we can obtain the much stronger (via the Chernoff bound) G(T):S exp(-O((a - T)2n)) , which is exponentially small in n. In the continuous case, if lGen,oo (x) has a density that, within some small interval [-m,mJ, is uniformly bounded by O(l/n), then we also have G(T) = O(T). In any case, we also have the following Corollary to Theorem 4. Corollary 6 Let the conditions of Theorem 4 hold, and suppose that G(T) :S Eo/2+ F(T) for some function F(T) (independent of n) that satisfies F(T) -+ 0 as T -+ 0, and some fixed EO > O. Then for ?(hGen) :S c(hGen,oo) + EO to hold with high pima (continuous) adult (continuous) 0.5 0.45 0.4 e ;;; 0.35 0.3 0.250 boston (predict it > median price, continuous) 0.5 0.45, - - - - - - - - - - - - - , 0.45 " 0.4 0.4 ~-2~~~,____ ~ 0.3 '--, 20 40 I::: \ ...~~ __ gO.35 0.25 0"0 60 optdigits (O's and 1 's, continuous) 10 20 02Q- - -""2"' 0- - -4"'0-----"cJ 60 30 optdigits (2's and 3's, continuous) ionosphere (continuous) o.4,---- - - - - - - - - - , 0 .4", - - - - - - - - - ----. 0.3 0 .3 0.5,---- - - - - - - - - - - - - - , ~0.2 01 ~ 0.1 0.2 50 100 150 200 adult (discrete) liver disorders (continuous) 0.7,---- - - - - - - - - - - - - - , 0.5, - - - - - - - - - - - - - - - - - , 0.6 0.45 ~ ~ 0.4 ~ 0.35 0.4 0.3 0.350 20 40 60 0.250 20 promoters (discrete) 60 80 100 100 120 lymphography (discrete) 0.5 200 300 400 breast cancer (discrete) 0.5 ~:: ~ ~ 40 ,......?...?..??.?.?..?.?..?.?.?.?.?...?...?. ?.. _. 0.2 ~ ~'\~.~:::.:~ ~0 . 3 0.4 "'" 0.45 ~ ..... 0.1 0 .2 ..... ".""""",,. 0.3 0.10'------,5~ 0---.,. 10~0---.---,J %L--~ 20~-~ 40---'-6~0---'---'8~0-~ 100 150 lenses (predict hard vs. soft, discrete) 0.250 0.4 0.8, - - - - - - - - - - - - - , 0.4 0'6 \=~_ 0.3 gOA ~ 0 .2 ..... 100 200 300 voting records (discrete) sick (discrete) 0.5 0.2 0.4 ~ 0.35 ?...? gO.2 ~ 0.1 \--. ------------ -- --- --- --- 0.10'---~-~10c--~ 15--2~0-~25 %~----,5~0---~10~0~--,,! 150 00 20 40 60 80 Figure 1: Results of 15 experiments on datasets from the VCI Machine Learning repository. Plots are of generalization error vs. m (averaged over 1000 random train/test splits). Dashed line is logistic regression; solid line is naive Bayes. probability, it suffices to pick m = O(log n). Note that the previous discussion implies that the preconditions of the Corollary do indeed hold in the case that the naive Bayes (and Proposition 5's) assumption holds , for any constant fa so long as n is large enough that fa ::::: exp( -O(o:2n)) (and similarly for the bounded Var(lGen ,oo (x)) case, with the more restrictive fa ::::: O(I/(o: 2n 17))). This also means that either ofthese (the latter also requiring T) > 0) is a sufficient condition for the asymptotic sample complexity to be 0 (log n). 4 Experiments The results of the previous section imply that even though the discriminative logistic regression algorithm has a lower asymptotic error, the generative naive Bayes classifier may also converge more quickly to its (higher) asymptotic error. Thus, as the number of training examples m is increased, one would expect generative naive Bayes to initially do better, but for discriminative logistic regression to eventually catch up to, and quite likely overtake, the performance of naive Bayes. To test these predictions, we performed experiments on 15 datasets, 8 with continuous inputs, 7 with discrete inputs, from the VCI Machine Learning repository.2 The results ofthese experiments are shown in Figure 1. We find that the theoretical predictions are borne out surprisingly well. There are a few cases in which logistic regression's performance did not catch up to that of naive Bayes, but this is observed primarily in particularly small datasets in which m presumably cannot grow large enough for us to observe the expected dominance of logistic regression in the large m limit. 5 Discussion Efron [2] also analyzed logistic regression and Normal Discriminant Analysis (for continuous inputs) , and concluded that the former was only asymptotically very slightly (1/3- 1/2 times) less statistically efficient. This is in marked contrast to our results, and one key difference is that, rather than assuming P(xly) is Gaussian with a diagonal covariance matrix (as we did), Efron considered the case where P(xly) is modeled as Gaussian with a full convariance matrix. In this setting, the estimated covariance matrix is singular if we have fewer than linear in n training examples, so it is no surprise that Normal Discriminant Analysis cannot learn much faster than logistic regression here. A second important difference is that Efron considered only the special case in which the P(xly) is truly Gaussian. Such an asymptotic comparison is not very useful in the general case, since the only possible conclusion, if ?(hDis,oo) < ?(hGen,oo), is that logistic regression is the superior algorithm. In contrast, as we saw previously, it is in the non-asymptotic case that the most interesting "two-regime" behavior is observed. Practical classification algorithms generally involve some form of regularization- in particular logistic regression can often be improved upon in practice by techniques 2To maximize the consistency with the theoretical discussion, these experiments avoided discrete/continuous hybrids by considering only the discrete or only the continuous-valued inputs for a dataset where necessary. Train/test splits were random subject to there being at least one example of each class in the training set, and continuous-valued inputs were also rescaled to [0 , 1] if necessary. In the case of linearly separable datasets, logistic regression makes no distinction between the many possible separating planes. In this setting we used an MCMC sampler to pick a classifier randomly from them (i.e., so the errors reported are empirical averages over the separating hyperplanes) . Our implementation of Normal Discriminant Analysis also used the (standard) trick of adding ? to the diagonal of the covariance matrix to ensure invertibility, and for naive Bayes we used I = 1. such as shrinking the parameters via an L1 constraint, imposing a margin constraint in the separable case, or various forms of averaging. Such regularization techniques can be viewed as changing the model family, however, and as such they are largely orthogonal to the analysis in this paper, which is based on examining particularly clear cases of Generative-Discriminative model pairings. By developing a clearer understanding of the conditions under which pure generative and discriminative approaches are most successful, we should be better able to design hybrid classifiers that enjoy the best properties of either across a wider range of conditions. Finally, while our discussion has focused on naive Bayes and logistic regression, it is straightforward to extend the analyses to several other models , including generativediscriminative pairs generated by using a fixed-structure , bounded fan-in Bayesian network model for P(xly) (of which naive Bayes is a special case). Acknowledgments We thank Andrew McCallum for helpful conversations. A. Ng is supported by a Microsoft Research fellowship. This work was also supported by a grant from Intel Corporation, NSF grant IIS-9988642, and ONR MURI N00014-00-1-0637. References [1] M. Anthony and P. Bartlett. Neural Network Learning: Th eoretical Foundations. Cambridge University Press, 1999. [2] B. Efron. The efficiency of logistic regression compared to Normal Discriminant Analysis. Journ. of the Amer. Statist. Assoc., 70:892- 898 , 1975. [3] P. Goldberg and M. Jerrum. Bounding the VC dimension of concept classes parameterized by real numbers. Machine Learning, 18:131-148, 1995. [4] G.J. McLachlan. Discriminant Analysis and Statistical Pattern Recognition. Wiley, New York, 1992. [5] Y. D. Rubinstein and T. Hastie. Discriminative vs. informative learning. In Proceedings of th e Third International Conference on Knowledge Discovery and Data Mining, pages 49- 53. AAAI Press, 1997. [6] V. N. Vapnik. Statistical Learning Theory. John Wiley & Sons, 1998.
2020 |@word repository:2 version:4 polynomial:1 seems:3 stronger:2 covariance:4 pick:3 solid:1 offering:1 must:2 john:1 informative:1 j1:1 plot:1 v:4 aside:1 generative:24 implying:1 fewer:1 plane:1 mccallum:1 record:1 provides:1 hyperplanes:1 ipi:1 direct:1 pairing:1 theoretically:1 expected:2 indeed:5 behavior:1 roughly:2 frequently:2 nor:1 tily:1 considering:1 becomes:1 begin:2 classifies:1 bounded:6 moreover:2 agnostic:1 what:2 corporation:1 nj:1 guarantee:2 berkeley:4 every:1 ti:2 voting:1 exactly:1 classifier:30 wrong:1 assoc:1 ly:11 enjoy:1 grant:2 positive:5 understood:1 limit:1 abuse:1 suggests:1 range:1 statistically:1 averaged:1 practical:1 acknowledgment:1 arguing:1 union:1 practice:1 empirical:4 significantly:1 cannot:2 close:1 optimize:2 equivalent:1 map:2 missing:1 straightforward:3 regardless:1 go:2 focused:1 simplicity:2 disorder:1 pure:1 rule:1 deriving:1 population:3 jli:1 coordinate:1 traditionally:1 laplace:1 suppose:2 us:1 ixi:1 goldberg:1 trick:1 recognition:1 particularly:2 predicts:2 muri:1 observed:2 worst:2 calculate:2 precondition:1 ensures:1 rescaled:1 intuition:2 complexity:4 depend:1 upon:1 division:1 efficiency:1 basis:2 completely:2 ogn:2 easily:2 joint:3 po:11 various:1 tx:1 llogp:1 articulated:1 train:2 distinct:2 approached:1 rubinstein:1 whose:1 quite:1 widely:1 solve:2 bey:2 say:2 larger:1 valued:2 jerrum:1 ip:1 rr:1 relevant:1 convergence:3 converges:4 wider:1 oo:54 andrew:2 depending:1 stat:1 liver:1 clearer:1 xly:8 implies:6 posit:1 correct:2 vc:5 everything:1 suffices:3 generalization:3 really:1 preliminary:1 convariance:1 proposition:12 summation:1 hold:8 considered:2 normal:6 exp:5 hois:9 presumably:1 predict:2 substituting:1 label:9 saw:1 mclachlan:1 always:2 gaussian:6 rather:5 corollary:3 focus:2 likelihood:3 contrast:2 helpful:1 initially:1 journ:1 issue:1 classification:5 overall:1 logn:2 misclassifies:1 prevailing:2 smoothing:2 fairly:2 special:2 never:1 ng:2 chernoff:4 t2:1 few:1 primarily:1 randomly:2 divergence:1 replaced:1 n1:1 microsoft:1 tlx:1 mining:1 truly:2 analyzed:1 upperbounded:1 held:1 necessary:2 folk:1 orthogonal:1 loosely:1 theoretical:3 increased:3 classify:1 modeling:1 compelling:1 typified:1 soft:1 uniform:1 examining:1 successful:1 reported:1 density:1 overtake:1 international:1 sequel:1 probabilistic:1 picking:1 michael:1 iy:3 quickly:1 again:2 aaai:1 possibly:2 borne:2 worse:1 li:1 invertibility:1 matter:1 piece:1 performed:1 view:1 picked:1 doing:1 observing:1 bayes:28 minimize:2 variance:3 largely:2 wisdom:1 bayesian:1 iid:1 definition:1 proof:2 gain:1 proved:2 dataset:1 recall:1 conversation:1 efron:4 knowledge:1 cj:1 actually:1 higher:4 improved:1 amer:1 though:5 strongly:1 lastly:1 correlation:1 sketch:2 ei:2 vci:2 logistic:28 requiring:1 true:5 concept:1 former:2 hence:1 regularization:2 conditionally:1 ll:1 criterion:2 tn:7 l1:1 goa:1 fi:2 superior:1 multinomial:1 empirically:1 exponentially:1 analog:2 extend:1 eoretical:1 refer:1 cambridge:1 imposing:1 consistency:1 similarly:5 add:1 sick:1 posterior:1 n00014:1 inequality:3 binary:2 onr:1 ey:1 eo:5 converge:1 maximize:2 tempting:1 dashed:1 ii:1 full:1 stem:1 technical:1 faster:4 long:4 ensuring:1 prediction:7 involving:1 regression:28 breast:1 expectation:3 metric:1 want:1 fellowship:1 interval:1 median:1 leaving:1 grow:1 concluded:1 singular:1 rest:1 induced:1 subject:1 contrary:1 spirit:1 jordan:1 call:1 near:1 intermediate:1 split:2 insofar:1 enough:2 fit:9 hastie:1 chebyshev:3 whether:1 bartlett:1 york:1 useful:1 generally:1 clear:1 involve:1 ylx:4 statist:1 nsf:1 eim:1 sign:1 estimated:1 lgen:29 correctly:1 discrete:17 dominance:1 key:2 terminology:1 drawn:1 changing:1 neither:1 asymptotically:2 fraction:4 sum:2 parameterized:2 almost:3 family:5 scaling:1 bound:10 guaranteed:1 fan:1 precisely:1 constraint:2 argument:2 separable:3 developing:1 according:2 smaller:2 slightly:2 across:1 son:1 ofthese:2 pr:4 computationally:1 equation:2 remains:1 previously:1 turn:1 count:1 loose:2 eventually:1 needed:2 letting:1 observe:1 away:1 occurrence:1 alternative:1 ho:1 shortly:1 cf:1 ensure:1 lat:1 upperbounds:1 restrictive:1 establish:1 already:1 quantity:2 occurs:1 parametric:2 fa:3 diagonal:3 div:1 thank:1 separating:2 extent:1 consensus:1 discriminant:10 reason:1 trivial:2 assuming:2 modeled:1 lg:1 pima:1 negative:2 stated:1 implementation:1 design:1 upper:1 observation:2 datasets:4 finite:1 n82:2 xily:4 perturbation:3 pair:6 required:1 kl:1 namely:1 generativediscriminative:1 california:2 distinction:1 adult:2 able:1 usually:1 pattern:1 regime:3 oj:1 including:2 belief:3 everyone:1 unrealistic:1 event:3 natural:1 difficulty:1 hybrid:2 indicator:1 imply:1 igen:1 picture:1 negativity:1 catch:2 naive:27 prior:1 understanding:1 discovery:1 asymptotic:17 expect:1 interesting:2 var:3 foundation:1 hgen:37 sufficient:2 thresholding:2 cancer:1 succinctly:1 surprisingly:1 supported:2 free:1 dis:4 formal:1 side:2 fall:2 taking:1 dimension:6 avoided:1 far:3 lymphography:1 ignore:1 preferred:2 implicitly:1 assumed:1 conclude:1 discriminative:31 xi:9 continuous:20 learn:3 mj:1 ca:2 ignoring:1 obtaining:1 anthony:1 did:2 promoter:1 linearly:2 bounding:1 prx:1 repeated:1 intel:1 en:1 wiley:2 shrinking:1 third:2 ix:1 theorem:4 specific:1 showing:3 ionosphere:1 vapnik:3 false:1 adding:1 conditioned:2 margin:1 xiiy:3 boston:1 surprise:1 logarithmic:2 likely:3 univariate:1 ijj:5 ldis:1 corresponds:2 chance:4 satisfies:1 conditional:1 optdigits:2 marked:1 viewed:1 price:1 hard:1 specifically:1 except:1 uniformly:2 sampler:1 averaging:1 lemma:3 lens:1 formally:1 support:1 latter:4 dept:1 mcmc:1 handling:1
1,120
2,021
Group Redundancy Measures Reveal Redundancy Reduction in the Auditory Pathway Gal Chechik Amir Globerson Naftali Tishby School of Computer Science and Engineering and The Interdisciplinary Center for Neural Computation Hebrew University of Jerusalem , Israel ggal@cs.huji.ac.il Michael J. Anderson Eric D. Young Department of Biomedical Engineering Johns Hopkins University, Baltimore, MD, USA Israel N elken Department of Physiology, Hadassah Medical School and The Interdisciplinary Center for Neural Computation Hebrew University of Jerusalem, Israel Abstract The way groups of auditory neurons interact to code acoustic information is investigated using an information theoretic approach. We develop measures of redundancy among groups of neurons, and apply them to the study of collaborative coding efficiency in two processing stations in the auditory pathway: the inferior colliculus (IC) and the primary auditory cortex (AI). Under two schemes for the coding of the acoustic content, acoustic segments coding and stimulus identity coding, we show differences both in information content and group redundancies between IC and AI neurons. These results provide for the first time a direct evidence for redundancy reduction along the ascending auditory pathway, as has been hypothesized for theoretical considerations [Barlow 1959,2001]. The redundancy effects under the single-spikes coding scheme are significant only for groups larger than ten cells, and cannot be revealed with the redundancy measures that use only pairs of cells. The results suggest that the auditory system transforms low level representations that contain redundancies due to the statistical structure of natural stimuli, into a representation in which cortical neurons extract rare and independent component of complex acoustic signals, that are useful for auditory scene analysis. 1 Introduction How do groups of sensory neurons interact to code information and how do these interactions change along the ascending sensory pathways? According to the a common view, sensory systems are composed of a series of processing stations, representing more and more complex aspects of sensory inputs. The changes in representations of stimuli along the sensory pathway reflect the information processing performed by the system. Several computational principles that govern these changes were suggested, such as information maximization and redundancy reduction [2, 3, 11]. In order to investigate such changes in practice, it is necessary to develop methods to quantify information content and redundancies among groups of neurons, and trace these measures along the sensory pathway. Interactions and high order correlations between neurons were mostly investigated within single brain areas on the level of pairs of cells (but also for larger groups of cells [9]) showing both synergistic and redundant interactions [8, 10, 21, 6, 7, 13]. The current study develops information theoretic redundancy measures for larger groups of neurons , focusing on the case of stimulus-conditioned independence. We then compare these measures in electro-physiological recordings from two auditory stations: the auditory mid-brain and the primary auditory cortex. 2 Redundancy measures for groups of neurons To investigate high order correlations and interactions within groups of neurons we start by defining information measures for groups of cells and then develop information redundancy measures for such groups. The properties of these measures are then further discussed for the specific case of stimulus-conditioned independence. Formally, the level of independence of two variables X and Y is commonly quantified by their mutual information (MI) [17,5]. This well known quantity, now widely used in analysis of neural data, is defined by J(X; Y) = DKL[P(X, Y)IIP(X)P(Y)] = ~p(x, y)log (:~~~~~)) (1) and measures how close the joint distribution P(X, Y) is to the factorization by the marginal distributions P(X)P(Y) (DKL is the Kullback Leiber divergence [5]). For larger groups of cells, an important generalized measure quantifies the information that several variables provide about each other. This multi information measure [18] is defined by (2) Similar to the mutual information case, the multi information measures how close the joint distribution is to the factorization by the marginals. It thus vanishes when variables are independent and is otherwise positive. We now turn to develop measures for group redundancies. Consider first the simple case of a pair of neurons (Xl, X 2 ) conveying information about the stimulus S. In this case, the redundancy-synergy index ([4, 7]) is defined by (3) Intuitively, RSpairs measures the amount of information on the stimulus S gained by observing the joint distribution of both Xl and X 2 , as compared with observing the two cells independently. In the extreme case where Xl = X 2 , the two cells are completely redundant and provide the same information about the stimulus, yielding RSpairs = I(Xl' X 2 ; S) - I(X l ; S) - I(X2 ; S) = -I(Xl; S), which is always non-positive. On the other hand, positive RSpairs values testify for synergistic interaction between Xl and X 2 ([8, 7, 4]). For larger groups of neurons, several different measures of redundancy-synergy may be considered, that encompass different levels of interactions. For example, one can quantify the residual information obtained from a group of N neurons compared to all its N - 1 subgroups. As with inclusion-exclusion calculations this measure takes the form of a telescopic sum: RSNIN-l = I(XN; S) - L{X N -l} I(X N -\ S) + ... + (_l)N-l L{Xd I(Xi ; S), where {Xk} are all the subgroups of size k out of the N available neurons. Unfortunately, this measure involves 2N information terms, making its calculation infeasible even for moderate N values 1. A different RS measure quantifies the information embodied in the joint distribution of N neurons compared to that provided by N single independent neurons, and is defined by N RSN ll = I(Xl ' ... , X N ; S) - 2..: I(Xi ; S) (4) i=l Interestingly, this synergy-redundancy measure may be rewritten as the difference between two multi-information terms N I(Xl ' ... , X N ; S) - 2..: I(Xi ; S) = (5) i= l N H(Xl' ... ,XN) - H(Xl' ... , XNIS) - 2..: H(Xi ) - H(XiIS) = i=l I(X l ; ... ; XNIS) - I(X l ; ... ;XN) where H(X) = - L x P(x)log(P(x)) is the entropy of X 2 . We conclude that the index RSN ll can be separated into two terms: one that is always non-negative, and measures the coding synergy, and the second which is always non-positive and quantifies the redundancy. These two terms correspond to two types of interactions between neurons: The first type are within-stimulus correlations (sometimes termed noise correlations) that emerge from functional connections between neurons and contribute to synergy. The second type are between stimulus correlations (or across stimulus correlations) that reflect the fact that the cells have similar responses per stimulus, and contribute to redundancy. Being interested in the latter type of correlations, we limit the discussion to the redundancy term -I(Xl; ... ; XN)' Formulating RSN ll as in equation 5 proves highly useful when neural activities are independent given the stimulus P(XIS) = II~l P(XiIS). In this case, the first (synergy) term vanishes , thus limiting neural interactions to the redundant lOur results below suggest that some redundancy effects become significant only for groups larger than 10-15 cells. 2When comparing redundancy in different processing stations, one must consider the effects of the baseline information conveyed by single neurons. We thus use the normalized redundancy (compare with [1 5] p.315 and [4]) defined by !iS Nll = RSNldI(Xl; ... ; X N ; S) regime. More importantly, under the independence assumption we only have to estimate the marginal distributions P(XiIS = s) for each stimulus s instead of the full distribution P(XIS = s). It thus allows to estimate an exponentially smaller number of parameters, which in our case of small sample sizes, provides more accurate information estimates. This approximation makes it possible to investigate redundancy among considerably larger groups of neurons than the 2-3 neuron groups considered previously in the literature. How reasonable is the conditional-independence approximation ? It is a good approximation whenever neuronal activity is mostly determined by the presented stimulus and to a lesser extent by interactions with nearby neurons. A possible example is the high input regime of cortical neurons receiving thousands of inputs, where a single input has only a limited influence on the activity of the target cell. The experimental evidence in this regard is however mixed (see e.g.[9]). One should note however, that stimulus-conditioned independence is implicitly assumed in analysis of non-simultaneously recorded data. To summarize, the stimulus-conditioned independence assumption limits interactions to the redundant regime, but allows to compare the extent of redundancy among large groups of cells in different brain areas. 3 Experimental Methods To investigate redundancy in the auditory pathway, we analyze extracellular recordings from two brain areas of gas-anesthetized cats: 16 cells from the Inferior Colliculus (Ie) - the third processing station of the ascending auditory pathway - and 19 cells from the Primary Auditory Cortex (AI) - the fifth station. Neural activity was recorded non-simultaneously from a total of 6 different animals responding to a set of complex natural and modified stimuli. Because cortical auditory neurons respond differently to simple and complex stimuli [12 , 1], we refrain from using artificial over-simplified acoustic stimuli but instead use a set of stimuli based on bird vocalizations which contains complex 'real-life' acoustic features. A representative example is shown in figure 1. 7 6 Q) "0 ."1i.e E '" 20 40 60 80 time (milliseconds) 100 20 40 60 80 100 time (milliseconds) Figure 1: A representative stimulus containing a short bird vocalization recorded in a natural environment. The set of stimuli consisted of similar natural and modified recordings. A. Signal in time domain B. Signal in frequency domain. 4 Experimental Results In practice, in order to estimate the information conveyed by neural activity from limited data, one must assume a decoding procedure, such as focusing on a simple statistic of the spike trains that encompasses some of its informative properties. In this paper we consider two extreme cases: coding short acoustic segments with single spikes and coding the stimulus identity with spike counts in a long window. In addition, we estimated information and redundancy obtained with two other statistics. First, the latency of the first spike after stimulus onset, and secondly, a statistic which generalizes the counts statistics for a general renewal process [19]. These calculations yielded higher information content on average, but similar redundancies as presented below. Their detailed results will be reported elsewhere. 0.15 1.2 ~ Inlerior Colliculus (IC) 0.1 c u'" ~0 . 8 C :::l .$ :0 U iO.6 ~ 0.05 <ii c o 0.4 t5 jg 0 Auditory Cortex (AI) o~========== 5 10 no of cells 15 -0.05 L--".--'--~---~--~--- o 5 10 no of cells 15 20 Figure 2: A. Information about stimulus frames as a function of number of cells. Information calculation was repeated for several subgroups of each size, and with several random seed initializations. The dark curve depicts the expected information provided by independent neurons (this expected curve is corrected for saturation effects [16] and is thus sub linear). The curved line depicts average information from joint distribution of sets of neurons Mean[J(Xl' ... Xk; S)]. All information estimations were corrected for small-samples bias by shuffling methods [14] . B. Fractional redundancy (difference of the mutual information from the expected baseline information divided by the baseline) as a function of number of neurons. 4.1 Coding acoustics with single spikes The current section focuses on the relation between single spikes and short windows of the acoustic stimuli shortly preceding them (which we denote as frames). As the set of possible frames is very large and no frame actually repeats itself, we must first pre-process the stimuli to reduce frames dimensionality. To this end, we first transformed the stimuli into the frequency domain (roughly approximating the cochlear transformation) and then extracted overlapping windows of 50 millisecond length, with 1 millisecond spacing. This set was clustered into 32 representatives, using a metric that groups together acoustic segments with the same spectro-temporal energy structure. This representation allowed us to estimate the joint distribution (under the stimulus-conditioned independence assumption) of cells' activity and stimuli, for groups of cells of different sizes. Figure 2A shows the mutual information between spikes and stimulus frames as a function of the number of cells for both AI and Ie neurons. Ie neurons convey high information but largely deviate from the information expected for independent neurons. On the other hand, AI neurons provide an order of magnitude less information than Ie cells but their information sums almost linearly, as expected from independent neurons. The difference between an information curve and its linear baseline measures the redundancy RSN II of equation 5. Figure 2B presents the normalized redundancy as a function of number of cells, showing that Ie cells are significantly more redundant than AI cells. 0.6 rr==--------,--------; D _ 0.6 rr==--------,--------; D Primary Auditory Cortex A 1 Inferior Colliculus IC _ 0.5 ~'-"'=-c.:..::..::.:..:.:::.'--===::...:..::-----" Primary Auditory Cortex A 1 Inferior Colliculus IC 0.5 ~'-"'=--"-"-':.:..:.:::.'--===::...:..::'---------" Q) Q) u c u c 504 504 u u u u o o 00.3 00.3 :?"' :?"' ~02 ~0.2 0. 0. o o 0.1 o 0.1 -0.5 -04 -0.3 -0.2 -0.1 pairwise redundancy -I(X ;Y)/I(X ;Y;S) o -8.8triplets fractional -0.6 -04 redundancy -0.2 0 -I(X;Y;Z)/I(X;Y;Z;S) Figure 3: Distribution of pairs (A.) and triplets (B.) normalized redundancies. AI cells (light bars) are significantly more independent than Ie cells (dark bars). Spike counts were collected over a window that maximizes mean single cells MI. Number of bins in counts-histogram was optimized separately for every cell. Information estimations were corrected for small-samples bias by shuffling methods [14]. 4.2 Coding stimuli by spike counts We now turn to investigate a second coding paradigm, and calculate the information conveyed by AI and Ie spike counts about the identity of the presented stimulus. To this end, we calculate a histogram of spike counts and estimate the counts' distribution as obtained from repeated presentations of the stimuli. The distribution of fractional redundancy in pairs of AI and Ie neurons is presented in figure 3A, and that of triplets in figure 3B 3 . As in the case of coding with single spikes, single AI cells convey on average less information about the stimulus. However, they are also more independent, thus making it possible to gain more information from groups of neurons. Ie neurons on the other hand, provide more information when considered separately but are more redundant. As in the case of coding acoustics with single spikes, single Ie cells provide more information than AI cells (data not shown) but this time AI cells convey half the information that Ie cells provide, while they convey ten times less information than Ie cells about acoustics. This suggests that AI cells poorly code the physical characteristics of the sound but convey information about its global properties. To illustrate the high information provided by both sets, we trained a neural network classifier that predicts the identity of the presented stimulus according to spike counts of a limited set of neurons. Figure 4 shows that both sets of neurons achieve considerable prediction accuracy, but Ie neurons obtain average accuracy of more than 90 percent already with five cells, while the average prediction accuracy using cortical neurons rises continuously 4. 3Unlike the binary case of single spikes, the limited amount of data prevents a robust estimation of information from spike counts for more than triplets of cells. 4The probability of accurate prediction is exponentially related to the input-output mutual information, via the relation Pcorrect = exp( -missing nats) yielding Mlnats = In(no. of stimuli) + In(Pcorrect). Classification thus provides lower bounds on information content . Figure 4. Prediction accuracy of stimulus identity as a function of number of Ie (upper curve) and AI (lower curve) cells used by the classifier. Error bars denote standard deviation across several subgroups of the same size. For each subgroup, a one-hidden layer neural network was trained separately for each stimulus using some stimulus presentations as a training set and the rest for testing. Performance reported is for the testing set. 5 0.95 >- "[IS "al 0.9 :0 <:0.85 o "u ~ 0.8 Q. 0.75 I II I I I?I II I II IIIII I I I I I jI" I 0.7 '-------=---~--~---~--~ 5 10 number of cells 15 20 Discussion We have developed information theoretic measures of redundancy among groups of neurons and applied them to investigate the collaborative coding efficiency in the auditory modality. Under two different coding paradigms, we show differences in both information content and group redundancies between Ie and cortical auditory neurons. Single Ie neurons carry more information about the presented stimulus, but are also more redundant. On the other hand, auditory cortical neurons carry less information but are more independent, thus allowing information to be summed almost linearly when considering groups of few tens of neurons. The results provide for the first time direct evidence for redundancy reduction along the ascending auditory pathway, as has been hypothesized by Barlow [2, 3]. The redundancy effects under the single-spikes coding paradigm are significant only for groups larger than ten cells, and cannot be revealed with the standard redundancy measures that use only pairs of cells. Our results suggest that transformations leading to redundancy reduction are not limited to low level sensory processing (aimed to reduce redundancy in input statistics) but are applied even at cortical sensory stations. We suggest that an essential experimental prerequisite to reveal these effects is the use of complex acoustic stimuli whose processing occurs at high level processing stations. The above findings are in agreement with the view that along the ascending sensory pathways, the number of neurons increase, their firing rates decrease , and neurons become tuned to more complex and independent features. Together, these suggest that the neural representation is mapped into a representation with higher effective dimensionality. Interestingly, recent advances in kernel-methods learning [20] have shown that nonlinear mapping into higher dimension and over-complete representations may be useful for learning of complex classifications. It is therefore possible that such mappings provide easier readout and more efficient learning in the brain. Acknowledgements This work supported in part by a Human Frontier Science Project (HFSP) grant RG 0133/1998 and by a grant from the Israeli Ministry of Science. References [1] O. Bar-Yosef and I. Nelken. Responses of neurons in cat primary auditory cortex to bird chirps: Effects of temporal and spectral context. J. Neuroscience, in press, 2001. [2] H.B. Barlow. Sensory mechanisms, the reduction of redundancy, and intelligence. In Mechanisation of thought processes, pages 535- 539. Her Majesty's stationary office, London , 1959. [3] H .B. Barlow. Redundancy reduction revisited. Network: Computation in neural systems, 12:241-253, 200l. [4] N . Brenner, S.P . Strong, R . Koberle, R. de Ruyter van Steveninck, and W . Bialek. Synergy in a neural code. Neural Computation, 13(7):1531, 2000. [5] T.M. Cover and J.A. Thomas. The elements of information theory. Plenum Press, New York, 1991. [6] Y. Dan, J.M. Alonso, W.M. Usrey, and R.C. Reid. Coding of visual information by precisely correlated spikes in the lateral geniculate nucleus. Nature Neuroscience, [7] I. Gat and N. Tishby. Synergy and redundancy among brain cells of behaving monkeys. In M.S. Kearns, S.A. Solla, and D.A.Cohn, editors, Advances in Neural Information Proceedings systems, volume 11, Cambridge, MA, 1999. MIT Press. [8] T.J. Gawne and B.J. Richmond. How independent are the messages carried by adjacent inferior temporal cortical neurons? J. Neurosci., 13(7):2758- 2771, 1993. [9] P.M. Gochin, M. Colombo, G. A. Dorfman, G.L. Gerstein, and C.G. Gross. Neural ensemble coding in inferior temporal cortex. J. Neurophysiol., 71:2325- 2337, 1994. 1(6):501- 507, 1998. [10] M. Meister. Multineural codes in retinal signaling. Proc. Natl. Acad. Sci., 93:609- 614, 1996. [11] J .P. Nadal, N. Brunei, and N . Parga. Nonlinear feedforward networks with stochastic outputs: infomax implies redundancy reduction. Network: Computation in neural systems, 9:207- 217, 1998. [12] I. Nelken, Y. Rotman, and O. Bar-Yosef. Specialization of the auditory system for the analysis of natural sounds. In J. Brugge and P.F. Poon, editors, Central Auditory Processing and Neural Modeling. Plenum, New York, 1997. [13] S. Nirenberg, S.M. Carcieri, A.L. Jacobs, and P.E. Latham. Retinal ganglion cells act largely as independent encoders. Nature, 411:698- 701, 200l. [14] LM . Optican , T.J. Gawne, B.J. Richmond, and P .J . Joseph. Unbiased measures of transmitted information and channel capacity from multivariate neuronal data. Bioi. Cyber, 65:305- 310, 1991. [15] E . T. Rolls and A. Treves . Neural Networks and Brain Function. Oxford Univ . Press, 1998. [16] I. Samengo. Independent neurons representing a fintie set of stimuli: dependence of the mutual information on the number of units sampled. Network: Comput. Neural Syst., 12:21- 31 , 200l. [17] C.E. Shanon. A mathematical theory of communication. The Bell systems technical journal, 27:379- 423,623- 656, 1948. [18] M. Studenty and J. Vejnarova. The multiinformation function as a tool for measuring stochastic dependence. In M.I. Jordan, editor, Learning in Graphical Models, pages 261-297. Dordrecht: Kluwer, 1998. [19] C. van Vreeswijk. Information trasmission with renewal neurons. In J.M. Bower, editor, Computational Neuroscience: Trends in Research. Elsevier Press, 200l. [20] V.N. Vapnik. The nature of statistical learning theory. Springer-Verlag, Berlin, 1995. [21] DK. Warland, P. Reinagel, and M. Meister. Decoding visual information from a population of retinal ganglion cells. J. Neurophysiol., 78:2336- 2350, 1997.
2021 |@word r:1 jacob:1 carry:2 reduction:8 series:1 contains:1 tuned:1 interestingly:2 optican:1 current:2 comparing:1 must:3 john:1 informative:1 stationary:1 half:1 intelligence:1 amir:1 xk:2 short:3 provides:2 contribute:2 revisited:1 five:1 mathematical:1 along:6 direct:2 become:2 pathway:10 dan:1 pairwise:1 expected:5 roughly:1 multi:3 brain:7 window:4 considering:1 provided:3 project:1 maximizes:1 israel:3 nadal:1 monkey:1 developed:1 finding:1 gal:1 transformation:2 temporal:4 every:1 act:1 xd:1 classifier:2 unit:1 medical:1 grant:2 dorfman:1 reid:1 positive:4 engineering:2 limit:2 io:1 acad:1 oxford:1 firing:1 bird:3 initialization:1 chirp:1 quantified:1 suggests:1 multiinformation:1 factorization:2 limited:5 steveninck:1 globerson:1 testing:2 practice:2 signaling:1 procedure:1 area:3 bell:1 physiology:1 significantly:2 thought:1 chechik:1 pre:1 carcieri:1 suggest:5 cannot:2 synergistic:2 close:2 ggal:1 influence:1 context:1 telescopic:1 center:2 missing:1 jerusalem:2 independently:1 reinagel:1 importantly:1 population:1 limiting:1 plenum:2 target:1 agreement:1 element:1 trend:1 predicts:1 leiber:1 thousand:1 calculate:2 readout:1 solla:1 decrease:1 gross:1 vanishes:2 govern:1 environment:1 nats:1 trained:2 segment:3 eric:1 efficiency:2 completely:1 mechanisation:1 neurophysiol:2 joint:6 differently:1 cat:2 train:1 separated:1 univ:1 effective:1 london:1 artificial:1 dordrecht:1 whose:1 larger:8 widely:1 otherwise:1 nirenberg:1 statistic:5 itself:1 vocalization:2 nll:1 rr:2 interaction:10 poon:1 poorly:1 achieve:1 illustrate:1 develop:4 ac:1 school:2 strong:1 c:1 involves:1 implies:1 quantify:2 stochastic:2 human:1 bin:1 clustered:1 gochin:1 secondly:1 frontier:1 considered:3 ic:5 exp:1 seed:1 mapping:2 lm:1 estimation:3 proc:1 geniculate:1 tool:1 mit:1 always:3 modified:2 colombo:1 office:1 focus:1 majesty:1 richmond:2 baseline:4 elsevier:1 hidden:1 relation:2 her:1 transformed:1 interested:1 among:6 classification:2 rsn:4 animal:1 renewal:2 summed:1 mutual:6 marginal:2 stimulus:44 develops:1 few:1 composed:1 simultaneously:2 divergence:1 testify:1 message:1 brunei:1 investigate:6 highly:1 extreme:2 yielding:2 light:1 natl:1 accurate:2 necessary:1 theoretical:1 modeling:1 cover:1 measuring:1 maximization:1 deviation:1 rare:1 tishby:2 reported:2 encoders:1 considerably:1 huji:1 ie:16 interdisciplinary:2 rotman:1 receiving:1 decoding:2 infomax:1 michael:1 together:2 hopkins:1 continuously:1 reflect:2 recorded:3 central:1 iip:1 containing:1 leading:1 syst:1 de:1 retinal:3 coding:18 onset:1 performed:1 view:2 observing:2 analyze:1 start:1 collaborative:2 il:1 accuracy:4 roll:1 largely:2 characteristic:1 ensemble:1 correspond:1 conveying:1 parga:1 gawne:2 whenever:1 energy:1 frequency:2 mi:2 gain:1 auditory:24 sampled:1 fractional:3 dimensionality:2 brugge:1 actually:1 focusing:2 higher:3 response:2 anderson:1 biomedical:1 correlation:7 hand:4 iiiii:1 nonlinear:2 overlapping:1 cohn:1 reveal:2 usa:1 effect:7 hypothesized:2 contain:1 normalized:3 barlow:4 consisted:1 unbiased:1 adjacent:1 ll:3 inferior:6 naftali:1 generalized:1 theoretic:3 complete:1 latham:1 percent:1 consideration:1 common:1 functional:1 physical:1 ji:1 exponentially:2 volume:1 discussed:1 kluwer:1 marginals:1 significant:3 cambridge:1 ai:15 shuffling:2 inclusion:1 jg:1 cortex:8 behaving:1 multivariate:1 hadassah:1 exclusion:1 recent:1 moderate:1 termed:1 verlag:1 binary:1 refrain:1 life:1 transmitted:1 ministry:1 preceding:1 paradigm:3 redundant:7 signal:3 ii:6 encompass:1 full:1 sound:2 technical:1 calculation:4 long:1 divided:1 dkl:2 prediction:4 metric:1 histogram:2 sometimes:1 kernel:1 cell:44 addition:1 separately:3 spacing:1 baltimore:1 modality:1 rest:1 unlike:1 recording:3 cyber:1 electro:1 jordan:1 revealed:2 feedforward:1 independence:8 reduce:2 lesser:1 specialization:1 york:2 useful:3 latency:1 detailed:1 aimed:1 transforms:1 amount:2 dark:2 mid:1 ten:4 vejnarova:1 millisecond:4 estimated:1 neuroscience:3 per:1 xii:3 group:28 redundancy:46 sum:2 colliculus:5 respond:1 almost:2 reasonable:1 gerstein:1 bound:1 layer:1 yielded:1 activity:6 precisely:1 x2:1 scene:1 nearby:1 aspect:1 formulating:1 extracellular:1 department:2 according:2 yosef:2 across:2 smaller:1 joseph:1 making:2 intuitively:1 equation:2 previously:1 turn:2 count:10 mechanism:1 vreeswijk:1 ascending:5 end:2 meister:2 available:1 generalizes:1 rewritten:1 prerequisite:1 apply:1 spectral:1 shortly:1 thomas:1 responding:1 graphical:1 warland:1 prof:1 approximating:1 already:1 quantity:1 spike:19 occurs:1 primary:6 dependence:2 md:1 bialek:1 mapped:1 lateral:1 sci:1 capacity:1 berlin:1 alonso:1 cochlear:1 extent:2 collected:1 code:5 length:1 index:2 hebrew:2 lour:1 mostly:2 unfortunately:1 trace:1 negative:1 rise:1 allowing:1 upper:1 neuron:50 curved:1 gas:1 defining:1 communication:1 frame:6 station:8 treves:1 pair:6 connection:1 optimized:1 acoustic:13 subgroup:5 israeli:1 suggested:1 bar:5 below:2 regime:3 summarize:1 encompasses:1 saturation:1 natural:5 residual:1 representing:2 scheme:2 carried:1 extract:1 embodied:1 koberle:1 deviate:1 literature:1 acknowledgement:1 mixed:1 nucleus:1 conveyed:3 principle:1 editor:4 elsewhere:1 repeat:1 supported:1 infeasible:1 bias:2 emerge:1 anesthetized:1 fifth:1 van:2 regard:1 curve:5 dimension:1 xn:4 cortical:8 t5:1 sensory:10 nelken:2 commonly:1 simplified:1 spectro:1 implicitly:1 kullback:1 synergy:8 global:1 conclude:1 assumed:1 xi:6 quantifies:3 triplet:4 nature:3 channel:1 robust:1 ruyter:1 interact:2 investigated:2 complex:8 domain:3 linearly:2 neurosci:1 noise:1 repeated:2 allowed:1 convey:5 usrey:1 neuronal:2 representative:3 depicts:2 sub:1 xl:13 comput:1 bower:1 third:1 young:1 specific:1 showing:2 dk:1 physiological:1 evidence:3 essential:1 vapnik:1 gained:1 gat:1 magnitude:1 conditioned:5 easier:1 entropy:1 rg:1 ganglion:2 visual:2 prevents:1 springer:1 extracted:1 ma:1 conditional:1 bioi:1 identity:5 presentation:2 brenner:1 content:6 change:4 considerable:1 determined:1 corrected:3 kearns:1 total:1 hfsp:1 experimental:4 formally:1 latter:1 correlated:1
1,121
2,022
Learning Lateral Interactions for Feature Binding and Sensory Segmentation Heiko Wersing HONDA R&D Europe GmbH Carl-Legien-Str.30, 63073 Offenbach/Main, Germany heiko.wersing@hre-ftr.f.rd.honda.co.jp Abstract We present a new approach to the supervised learning of lateral interactions for the competitive layer model (CLM) dynamic feature binding architecture. The method is based on consistency conditions, which were recently shown to characterize the attractor states of this linear threshold recurrent network. For a given set of training examples the learning problem is formulated as a convex quadratic optimization problem in the lateral interaction weights. An efficient dimension reduction of the learning problem can be achieved by using a linear superposition of basis interactions. We show the successful application of the method to a medical image segmentation problem of fluorescence microscope cell images. 1 Introduction Feature binding has been proposed to provide elegant solution strategies to the segmentation problem in perception [11, 12, 14]. A lot of feature binding models have thus tried to reproduce groping mechanisms like the Gestalt laws of visual perception, e.g. connectedness and good continuation, using temporal synchronization [12] or spatial coactivation [9, 14] for binding. Quite generally in these models, grouping is based on lateral interactions between feature-representing neurons, which characterize the degree of compatibility between features. Currently in most of the approaches this lateral interaction scheme is chosen heuristically, since the experimental data on the corresponding connection patterns in the visual cortex is insufficient. Nevertheless, in more complex feature spaces this heuristic approach becomes infeasible, raising the question for more systematic learning methods for lateral interactions. Mozer et al. [4] suggested supervised learning for a dynamic feature binding model of complex-valued directional units, where the connections to hidden units guiding the grouping dynamics were adapted by recurrent backpropagation learning. The application was limited to synthetic rectangle patterns. Hofmann et al. [2] considered unsupervised texture segmentation by a pairwise clustering approach on feature vectors derived from Gabor filter banks at different frequencies and orientations. In their model the pairwise feature compatibilities are determined by a divergence measure of the local feature distributions which was shown to achieve good segmentation results for a range of image types. The problem of segmentation can also be phrased as a labeling problem, where relaxation labeling algorithms have been used as a popular tool in a wide range of computer vision applications. Pelillo & Refice [7] suggested a supervised learning method for the compatibility coefficients of relaxation labeling algorithms, based on minimizing the distance between a target labeling vector and the output after iterating a fixed number of relaxation steps. The main problem are multiple local minima arising in this highly nonlinear optimization problem. Recent results have shown that linear threshold (LT) networks provide interesting architectures for combining properties of digital selection and analogue context-sensitive amplification [1, 13] with efficient hardware implementation options [1]. Xie et al. [16] demonstrated how these properties can be used to learn winner-take-all competition between groups of neurons in an LT network with lateral inhibition. The CLM binding model is implemented by a large-scale topographically organized LT network, and it was shown that this leads to consistency conditions characterizing its binding states [14]. In this contribution we show how these conditions can be used to formulate a learning approach for the CLM as a quadratic optimization problem. In Section 2 we briefly introduce the competitive layer binding model. Our learning approach is elaborated in Section 3. In Section 4 we show application results of the approach to a cell segmentation problem and give a discussion in the final Section 5. 2 The CLM architecture The CLM [9, 14] consists of a set of layers of feature-selective neurons (see Fig. 1). The activity of a neuron at position  in layer  is denoted by  , and a column  denotes the set of the neuron activities   ,    , sharing a common position  . With each column a particular ?feature? is associated, which is described by a set of parameters like e.g. local edge elements characterized by position and orientation     . A binding between two features, represented by columns  and   , is expressed by simultaneous activities    ! and  "#  ! that share a common layer $ . All neurons in a column  are equally driven by an external input %& , which represents the significance of the detection of feature  by a preprocessing step. The afferent input %& is fed to the activities & with a connection  weight ' ( . Within each layer  the activities are coupled via lateral connections )  " which characterize the degree of compatibility between features  and   and which is a symmetric function of the feature parameters, thus )  " *)  "+ . The purpose of the layered arrangement in the CLM is to enforce an assignment of the input features to the layers by the dynamics, using the contextual information stored in the lateral interactions. The unique assignment to a single layer is realized by a columnar Winner-Take-All (WTA) circuit, which uses mutual symmetric inhibitory interactions with absolute strength ' , between neural activities - and ./ that share a common column  . Due to the WTA coupling, for a stable equilibrium state of the CLM only a neuron from one layer can be active within each column [14]. The number of layers does not predetermine the number of active groups, since for sufficiently many layers only those are active that carry a salient group. The combination of afferent inputs and lateral and vertical interactions is combined into the standard linear threshold additive activity dynamics 0 .1324&6587:9 ';<%.=2?> / &@/.5A> " )   " . "BDCE (1) where 7F&G!HJIK - . For ' large compared to the lateral weights )   " , the single active neuron in a column reproduces its afferent input,  ML %  . As was shown [14], the stable states of (1) satisfy the consistency conditions > " / ) "  "N/O > " )  "  "P for all  RQTU S G$ VW (2) which express the assignment of a feature  to the layer = $ V with highest lateral support. xrL xr?L layer L vertical WTA interaction xr2 lateral interaction xr?2 layer 2 xr1 xr?1 layer 1 hr hr? input Figure 1: The competitive layer model architecture (see text for description). 3 Learning of CLM Lateral Interactions Formulation of the Learning Problem. The overall task of the learning algorithm is  to adapt the lateral interactions, given by the interaction coefficients )  " , such that the CLM architecture performs appropriate segmentation on the labeled training data and also generalizes to new test data. We assume that the training data consists of a set of labeled training patterns ,   , where each pattern consists of a subset of different features with their corresponding labels $   . For     each labeled training pattern a target labeling vector is constructed by choosing                 Q,U for all  S $  V         (3)    /  for the labeled columns, assuming %   . Columns for features which are not contained   . In for all     S in the training pattern are filled with zeroes according to     the following indices     run over all possible  features, e.g. all edges of different orientations at different image positions, while  V run over the subset of features realized in a particular pattern, e.g. only one oriented edge at each image position. The assignment vectors  =  form the basis of the learning approach since they represent the target activity distribution, which we want to obtain after iterating the CLM with appropriately adjusted lateral interactions. In the following the abbreviation $ for $ V is used to keep the notation readable.     The goal of the learning process is to make the training patterns consistent, which is in accordance with (2) expressed by the inequalities > " / ) "    " / OT> "!     " )   "   "  for all #     @Q! S $  (4) These  2A  inequalities define the learning problem that we want to solve in the following. Let us develop a more compact notation. We can rewrite (4) as > & $#  % / ) %#  " " O! for all #    @QTU S G$ (5) #$% " & @/ where #  ' " )(  +* ( / # + " / 2,(  # + "P" - . The form of the inequalities can be simplified by introducing multiindices . and / which correspond to /10 32E     , .40 5#  RQE and # &98 & #$ %/ . The index . runs over all  2  !  consistency relations de)5670 ) % " , 6 0 "  8 with components fined for the labeled columns of the assignment vectors. The vectors : &;6 8 are called consistency vectors and represent the consistency constraints for the lateral interaction. The index / runs over all entries in the lateral interaction matrix. The vector   )   @)    )  )  with  components contains the corresponding matrix entries. The inequalities (4) can then be written in the form > & 6 8 )56 6 O .F for all (6) This illustrates the nature of the learning problem. The problem is to find a weight vector which leads to a lateral interaction matrix, such that the consistency vectors lie in the opposite half space of the weight state space. Since the conditions (6) determine the attractivity to achieve of the training patterns, it is customary to introduce a positive margin   greater robustness. This gives the target inequalities > & 6 8 )5665 6 O! .F for all (7) which we want to solve in for given training data. If the system of inequalities admits a solution for it is called compatible. If there is no satisfying all constraints, the system is called incompatible.  % #  Superposition of Basis Interactions. If the number of features is large, the number of parameters in the complete interaction matrix ) " may be too large to be robustly estimated from a limited number of training examples. To achieve generalization from the training data, it is necessary to reduce the number of parameters which have to be adapted during learning. This is also useful to incorporate a priori knowledge into the interaction. An example is to choose basis functions which incorporate invariances such as translation and rotation invariance, or which satisfy the constraint that the interaction is equal in all layers. A simple but powerful approach is to choose a set of  fixed basis interactions  with compatibilities  " J W  , with an interaction ) " obtained by linear superposition %#  ' #  # ) % " *>     % # "  >    6 (8) with weight coefficients  W  . Now the learning problem of solving the inequalities (7) can be recast in the new free parameters . After inserting (8) into (7) we obtain the transformed problem & 68    6   8 (9) 5 OA for all F . 6   8 &8  8 in the basis interacwhere  1! 6 6  6 is the component of the consistency vector :  tion  . The basis interactions can thus be used to reduce the dimensionality of the learning > > 51 > problem. To avoid any redundancy, the basis interactions should be linearly independent. Although the functions are here denoted ?basis? functions, they need neither be orthogonal  . nor span the whole space of interactions   Quadratic Consistency Optimization. The generic case in any real world application is that the majority of training vectors contains relevant information, while single spurious vectors may be present due to noise or other disturbing factors. Consequently, in most applications the equations (7) or (9) will be incompatible and can only be satisfied approximately. This will be especially the case, if a low-dimensional embedding is used for the basis function templates as described above. We therefore suggest to adapt the interactions by minimizing the following convex cost function  QCO  > 8 9 > 6 & 68 ) 6 5 C   (10) A similar minimization approach was suggested for the imprinting of attractors for the Brain-State-in-a-Box (BSB) model [8], and a recent study has shown that the approach is competitive with other methods for designing BSB associative memories [6]. For a fixed positive margin  A , the cost function (10) is minimized by making the inner products of the weight vector and the consistency vectors negative. The global minimum  is attained if the inner products are all equal to 2  , which can be interpreted with QCO  such that all consistency inequalities are fulfilled in an equal manner. Although this additional regularizing constraint is hard to justify on theoretical grounds, the later application shows that it works quite well for the application examples considered. If we insert the expansion of in the basis of function templates we obtain according to (8)  QCO  > 9 > 8   8 5 C  (11)  which results in a  -dimensional convex quadratic minimization problem in the parameters. The coefficients , which give the components of the training patterns in the basis @/     interactions, are given by  "  "N/  " 2  "  "  " . The quadratic optimization problem is then given by minimizing 8  8  ! 6 & 6 8 6 !   !       5 >    5 (12)     8 8 8 where    ! 8   and  ! 8  . If the coefficients  are unconstrained, then by solving the linear system of equations    the  minimum  !     of 5 (12)  can beforobtained all .  QCO   >    4 Application to Cell Segmentation The automatic detection and segmentation of individual cells in fluorescence micrographs is a key technology for high-throughput analysis of immune cell surface proteins [5]. The strong shape variability of cells in tissue, however, poses a strong challenge to any automatic recognition approach. Figure 2a shows corresponding fluorescence microscopy images from a tissue section containing lymphocyte cells (courtesy W. Schubert). In the bottom row corresponding image patches are displayed, where individual cell regions were manually labeled to obtain training data for the learning process.     For each of the image patches, a training vector consists of a list of labeled edge features parameterized by  E  , where E is the position in the image and E is a unit local edge orientation vector computed from the intensity gradient. For a pixel image this amounts to a set of  labeled edge features. Since the figure-ground separating mechanism as implemented by the CLM [14] is also used for this cell segmentation application, features which are not labeled as part of a cell obtain the corresponding background label, given by   . Each training pattern contains one additional free layer, to enable the learning algorithm to generalize over the number of layers. The lateral interaction to be adapted is decomposed into the following weighted basis components: i) A constant negative interaction between all features, which facilitates group separation, ii) a self-coupling interaction in the background layer which determines the attractivity of the background for figure-ground segmentation, and iii) an angular interaction with limited range, which is in itself decomposed into templates, capturing the interaction for a particular combination of the relative angles between two edges. This angular decomposition is done using a discretization of the space of orientations, turning the unit-vector representation into an angular orientation variable  . To achieve rotation invariance of the interaction, it is only dependent on the edge orientations relative    6 8 7 8 9 7 6 5 4 4 2 3 5 2 3 a) Manually labelled training patterns b) Grouping results after learning Figure 2: a) Original images and manually labeled training patterns from a fluorescence micrograph. b) Test patterns and resulting CLM segmentation with learned lateral interaction. Grayscale represents different layer activations, where a total of 20 layers plus one background layer (black) was used.   2 to their mutual position difference vector  . The angles  and  are dis  cretized by partitioning the interval into 8 subintervals. For each combination of the two discretized edge orientations there is an interaction template generated, which is only responding in this combined orientation interval. Thus the angular templates do not overlap in the combined     space, i.e. if    W     for a particular    , then   P         for all  S . Since the interaction must be symmetric under feature exchange, this does not result in   different combinations, but only 36 independent templates. Apart form the discretization, the interaction represents the most arbitrary angular-dependent interaction within the local neighborhood, which is symmetric under feature exchange. We use two sets of angular templates for O  and  O O  respectively, where  is the maximal local interaction range. With the abovementioned two components, the resulting optimization problem is 36+36+2=74dimensional. Figure 3 compares the optimized interaction field to earlier heuristic lateral interactions for contour grouping. See [15] for a more detailed discussion.  #  #           The performance of the learning approach was investigated by choosing a small number of the manually labeled patterns as training patterns. For all the training examples we used, the resulting inequalities (9) were in fact incompatible, rendering a direct solution of (9) infeasible. After training was completed by minimizing (12), a new image patch was selected as a test pattern and the CLM grouping was performed with the lateral interaction learned before, using the dynamical model as described in [14]. The quadratic consistency optimization was performed as described in the previous section, exploring the free margin parameter  . For a set of two training patterns as shown in Fig. (2)a with a total of 1600 features each, a learning sweep takes about 4 minutes on a standard desktop computer. Typical segmentation results obtained with the quadratic consistency optimization approach are shown in Figure 2b, where the margin was given by * . The grouping results were not very sensitive to  in a range of O  O . The grouping results show a good segmentation performance where most of the salient cells are detected as single groups. There are some spurious groups where a dark image region forms an additional group and some smaller cells are rejected into the background layer. Apart from these minor errors, the optimization has achieved an adequate balancing of the different lateral interaction components for this segmentation task.  n2 n1 n2 n2 d n1 p 1 p 2 b) Edge parameters a) Plotting scheme c) Standard continuity interaction field d) Learned interaction field Figure 3: Comparison between heuristic continuity grouping interaction field and a learned lateral interaction field for cell segmentation. The interaction depends on the difference vector and two unit vectors , shown in b), encoding directed orientation. a) ex plains the interaction visualizations c) and d) by showing a magnification of the plot c) of the interaction field of a single horizontal edge pointing to the left. The plots are generated by computing the interaction of the central directed edge with directed edges of all directions (like a cylindrical plot) at a spatial grid. Black edges share excitatory, white edges share inhibitory interaction with the central edge and length codes for interaction strength. The cocircular continuity field in c) depends on position and orientation but is not direction-selective. It supports pairs of edges which are cocircular, i.e. lie tangentially to a common circle and has been recently used for contour segmentation [3, 14]. The learned lateral interaction field is shown in d). It is direction-selective and supports pairs of edges which ?turn right?. The strong local support is balanced by similarly strong long-range inhibition. 5 Discussion The presented results show that appropriate lateral interactions can be obtained for the CLM binding architecture from the quadratic consistency optimization approach. The only a priori conditions which were used for the template design were the properties of locality, symmetry, and translation as well as rotation invariance. This supervised learning approach has clear advantages over the manual tuning of complex feature interactions in complex feature spaces with many parameters. We consider this as an important step towards practical applicability of the feature binding concept. The presented quadratic consistency optimization method is based on choosing equal margins for all consistency inequalities. There exist other approaches to large margin classifica- tion, like support vector machines [10], where more sophisticated methods were suggested for appropriate margin determination. The application of similar methods to the supervised learning of CLM interactions provides an interesting field for future work. Acknowledgments: This work was supported by DFG grant GK-231 and carried out at the Faculty of Technology, University of Bielefeld. The author thanks Helge Ritter and Tim Nattkemper for discussions and Walter Schubert for providing the cell image data. References [1] R. Hahnloser, R. Sarpeshkar, M. A. Mahowald, R. J. Douglas, and H. S. Seung. Digital selection and analogue amplification coexist in a cortex-inspired silicon circuit. Nature, 405:947?951, 2000. [2] T. Hofmann, J. Puzicha, and J. Buhmann. Unsupervised texture segmentation in a deterministic annealing framework. IEEE Trans. Pattern Analysis and Machine Intelligence, 20(8):803?818, 1998. [3] Z. Li. A neural model of contour integration in the primary visual cortex. Neural Computation, 10:903?940, 1998. [4] M. Mozer, R. S. Zemel, M. Behrmann, and C. K. I. Williams. Learning to segment images using dynamic feature binding. Neural Computation, 4(5):650?665, 1992. [5] T. W. Nattkemper, H. Ritter, and W. Schubert. A neural classificator enabling high-throughput topological analysis of lymphocytes in tissue sections. IEEE Trans. Inf. Techn. in Biomed., 5(2):138?149, 2001. [6] J. Park, H. Cho, and D. Park. On the design of BSB associative memories using semidefinite programming. Neural Computation, 11:1985?1994, 1999. [7] M. Pelillo and M Refice. Learning compatibility coefficients for relaxation labeling processes. IEEE Trans. Pattern Analysis and Machine Intelligence, 16(9):933?945, 1994. [8] Renzo Perfetti. A synthesis procedure for Brain-State-in-a-Box neural networks. IEEE Transactions on Neural Networks, 6(5):1071?1080, September 1995. [9] H. Ritter. A spatial approach to feature linking. In Proc. International Neural Network Conference Paris Vol.2, pages 898?901, 1990. [10] V. Vapnik. The nature of statistical learning theory. Springer, New York, 1995. [11] C. von der Malsburg. The what and why of binding: The modeler?s perspective. Neuron, 24:95?104, 1999. [12] D. Wang and D. Terman. Image segmentation based on oscillatory correlation. Neural Computation, 9(4):805?836, 1997. [13] H. Wersing, W.-J. Beyn, and H. Ritter. Dynamical stability conditions for recurrent neural networks with unsaturating piecewise linear transfer functions. Neural Computation, 13(8):1811? 1825, 2001. [14] H. Wersing, J. J. Steil, and H. Ritter. A competitive layer model for feature binding and sensory segmentation. Neural Computation, 13(2):357?387, 2001. [15] Heiko Wersing. Spatial Feature Binding and Learning in Competitive Neural Layer Architectures. PhD thesis, University of Bielefeld, 2000. Published by Cuvillier, Goettingen. [16] X. Xie, R. Hahnloser, and H.S. Seung. Learning winner-take-all competition between groups of neurons in lateral inhibition networks. In Advances in Neural Information Processing Systems, volume 13. The MIT Press, 2001.
2022 |@word cylindrical:1 faculty:1 briefly:1 heuristically:1 tried:1 decomposition:1 carry:1 reduction:1 contains:3 contextual:1 discretization:2 activation:1 written:1 must:1 additive:1 shape:1 hofmann:2 plot:3 half:1 selected:1 intelligence:2 desktop:1 provides:1 honda:2 constructed:1 direct:1 consists:4 manner:1 introduce:2 pairwise:2 nor:1 brain:2 discretized:1 inspired:1 decomposed:2 str:1 becomes:1 notation:2 circuit:2 what:1 interpreted:1 temporal:1 partitioning:1 unit:5 medical:1 grant:1 positive:2 before:1 local:7 accordance:1 encoding:1 micrographs:1 approximately:1 connectedness:1 black:2 plus:1 co:1 limited:3 range:6 coactivation:1 directed:3 unique:1 practical:1 acknowledgment:1 bsb:3 imprinting:1 backpropagation:1 xr:3 procedure:1 gabor:1 suggest:1 protein:1 selection:2 layered:1 coexist:1 context:1 deterministic:1 demonstrated:1 courtesy:1 williams:1 convex:3 formulate:1 embedding:1 stability:1 target:4 programming:1 carl:1 us:1 designing:1 element:1 satisfying:1 recognition:1 magnification:1 labeled:11 bottom:1 wang:1 region:2 highest:1 balanced:1 mozer:2 seung:2 dynamic:6 rewrite:1 solving:2 segment:1 topographically:1 basis:13 represented:1 sarpeshkar:1 walter:1 detected:1 zemel:1 labeling:6 choosing:3 neighborhood:1 quite:2 heuristic:3 valued:1 solve:2 itself:1 final:1 associative:2 advantage:1 interaction:60 product:2 maximal:1 inserting:1 relevant:1 combining:1 achieve:4 amplification:2 description:1 competition:2 tim:1 coupling:2 recurrent:3 develop:1 pose:1 minor:1 qt:1 pelillo:2 strong:4 implemented:2 direction:3 filter:1 enable:1 exchange:2 generalization:1 goettingen:1 adjusted:1 insert:1 exploring:1 sufficiently:1 considered:2 ground:3 equilibrium:1 pointing:1 purpose:1 proc:1 label:2 currently:1 superposition:3 fluorescence:4 sensitive:2 tool:1 weighted:1 minimization:2 mit:1 heiko:3 avoid:1 derived:1 dependent:2 hidden:1 relation:1 spurious:2 reproduce:1 selective:3 transformed:1 germany:1 biomed:1 compatibility:6 overall:1 schubert:3 orientation:11 pixel:1 denoted:2 priori:2 spatial:4 integration:1 mutual:2 equal:4 field:9 manually:4 represents:3 park:2 unsupervised:2 throughput:2 future:1 minimized:1 terman:1 piecewise:1 oriented:1 divergence:1 individual:2 dfg:1 attractor:2 n1:2 detection:2 highly:1 semidefinite:1 clm:15 edge:18 necessary:1 orthogonal:1 filled:1 circle:1 theoretical:1 column:10 earlier:1 assignment:5 mahowald:1 cost:2 introducing:1 applicability:1 subset:2 entry:2 lymphocyte:2 successful:1 too:1 characterize:3 stored:1 refice:2 synthetic:1 combined:3 cho:1 thanks:1 international:1 systematic:1 ritter:5 synthesis:1 von:1 central:2 satisfied:1 thesis:1 containing:1 choose:2 external:1 li:1 de:1 coefficient:6 satisfy:2 nattkemper:2 afferent:3 depends:2 tion:2 later:1 lot:1 performed:2 competitive:6 option:1 elaborated:1 contribution:1 tangentially:1 correspond:1 directional:1 generalize:1 published:1 tissue:3 simultaneous:1 oscillatory:1 sharing:1 manual:1 frequency:1 associated:1 modeler:1 popular:1 knowledge:1 dce:1 dimensionality:1 segmentation:21 organized:1 sophisticated:1 attained:1 supervised:5 xie:2 classifica:1 formulation:1 done:1 box:2 angular:6 rejected:1 correlation:1 horizontal:1 nonlinear:1 continuity:3 concept:1 symmetric:4 white:1 during:1 self:1 complete:1 performs:1 image:16 recently:2 steil:1 common:4 rotation:3 winner:3 jp:1 volume:1 linking:1 silicon:1 rd:1 unconstrained:1 consistency:16 automatic:2 grid:1 similarly:1 tuning:1 immune:1 stable:2 europe:1 cortex:3 surface:1 inhibition:3 recent:2 perspective:1 inf:1 driven:1 apart:2 inequality:10 der:1 minimum:3 greater:1 additional:3 attractivity:2 determine:1 ii:1 multiple:1 characterized:1 adapt:2 determination:1 fined:1 long:1 equally:1 predetermine:1 ftr:1 vision:1 represent:2 achieved:2 cell:14 microscope:1 microscopy:1 background:5 want:3 interval:2 annealing:1 appropriately:1 ot:1 elegant:1 facilitates:1 iii:1 rendering:1 architecture:7 opposite:1 reduce:2 inner:2 york:1 adequate:1 generally:1 iterating:2 useful:1 detailed:1 clear:1 xr2:1 amount:1 dark:1 hardware:1 continuation:1 exist:1 inhibitory:2 xr1:1 estimated:1 arising:1 fulfilled:1 hji:1 vol:1 express:1 group:8 redundancy:1 salient:2 key:1 threshold:3 nevertheless:1 micrograph:1 neither:1 douglas:1 rectangle:1 relaxation:4 run:4 angle:2 parameterized:1 powerful:1 bielefeld:2 patch:3 separation:1 incompatible:3 capturing:1 layer:26 quadratic:9 topological:1 activity:8 adapted:3 strength:2 constraint:4 phrased:1 span:1 according:2 combination:4 smaller:1 helge:1 wta:3 making:1 equation:2 visualization:1 turn:1 mechanism:2 fed:1 generalizes:1 enforce:1 appropriate:3 generic:1 robustly:1 robustness:1 customary:1 original:1 denotes:1 clustering:1 responding:1 completed:1 readable:1 malsburg:1 rqt:1 especially:1 sweep:1 question:1 arrangement:1 realized:2 strategy:1 primary:1 abovementioned:1 september:1 gradient:1 distance:1 lateral:28 oa:1 majority:1 separating:1 assuming:1 length:1 code:1 index:3 insufficient:1 providing:1 minimizing:4 gk:1 negative:2 implementation:1 design:2 vertical:2 neuron:10 enabling:1 displayed:1 variability:1 arbitrary:1 intensity:1 pair:2 paris:1 connection:4 optimized:1 raising:1 learned:5 trans:3 suggested:4 dynamical:2 perception:2 pattern:20 challenge:1 recast:1 memory:2 analogue:2 overlap:1 turning:1 hr:2 buhmann:1 representing:1 scheme:2 technology:2 carried:1 coupled:1 text:1 relative:2 law:1 synchronization:1 interesting:2 digital:2 degree:2 consistent:1 plotting:1 bank:1 share:4 balancing:1 translation:2 row:1 compatible:1 excitatory:1 supported:1 free:3 infeasible:2 dis:1 cocircular:2 wide:1 template:8 characterizing:1 absolute:1 dimension:1 plain:1 world:1 contour:3 sensory:2 author:1 disturbing:1 preprocessing:1 simplified:1 gestalt:1 transaction:1 compact:1 keep:1 ml:1 reproduces:1 active:4 global:1 grayscale:1 why:1 learn:1 nature:3 transfer:1 subintervals:1 symmetry:1 expansion:1 investigated:1 complex:4 significance:1 main:2 linearly:1 whole:1 noise:1 n2:3 gmbh:1 fig:2 position:8 guiding:1 lie:2 behrmann:1 minute:1 showing:1 list:1 admits:1 grouping:8 vapnik:1 texture:2 phd:1 illustrates:1 margin:7 columnar:1 locality:1 lt:3 visual:3 expressed:2 contained:1 binding:16 springer:1 determines:1 abbreviation:1 hahnloser:2 goal:1 formulated:1 consequently:1 towards:1 labelled:1 hard:1 wersing:5 determined:1 typical:1 justify:1 hre:1 called:3 total:2 techn:1 invariance:4 experimental:1 puzicha:1 support:5 incorporate:2 regularizing:1 ex:1
1,122
2,023
Pranking with Ranking Koby Crammer and Yoram Singer School of Computer Science & Engineering The Hebrew University, Jerusalem 91904, Israel {kobics,singer}@cs.huji.ac.il Abstract We discuss the problem of ranking instances. In our framework each instance is associated with a rank or a rating, which is an integer from 1 to k. Our goal is to find a rank-prediction rule that assigns each instance a rank which is as close as possible to the instance's true rank. We describe a simple and efficient online algorithm, analyze its performance in the mistake bound model, and prove its correctness. We describe two sets of experiments, with synthetic data and with the EachMovie dataset for collaborative filtering. In the experiments we performed, our algorithm outperforms online algorithms for regression and classification applied to ranking. 1 Introduction The ranking problem we discuss in this paper shares common properties with both classification and regression problems. As in classification problems the goal is to assign one of k possible labels to a new instance. Similar to regression problems, the set of k labels is structured as there is a total order relation between the labels. We refer to the labels as ranks and without loss of generality assume that the ranks constitute the set {I, 2, .. . , k} . Settings in which it is natural to rank or rate instances rather than classify are common in tasks such as information retrieval and collaborative filtering. We use the latter as our running example. In collaborative filtering the goal is to predict a user's rating on new items such as books or movies given the user's past ratings of the similar items. The goal is to determine whether a movie fan will like a new movie and to what degree, which is expressed as a rank. An example for possible ratings might be, run-to-see , very-good, good, only-if-you-must, and do-not-bother. While the different ratings carry meaningful semantics, from a learning-theoretic point of view we model the ratings as a totally ordered set (whose size is 5 in the example above). The interest in ordering or ranking of objects is by no means new and is still the source of ongoing research in many fields such mathematical economics, social science, and computer science. Due to lack of space we clearly cannot cover thoroughly previous work related to ranking. For a short overview from a learning-theoretic point of view see [1] and the references therein. One of the main results of [1] underscores a complexity gap between classification learning and ranking learning. To sidestep the inherent intractability problems of ranking learning several approaches have been suggested. One possible approach is to cast a ranking problem as a regression problem. Another approach is to reduce a total order into a set of pref- Correct interval #l \ I Figure 1: An Illustration of the update rule. erences over pairs [3, 5]. The first case imposes a metric on the set of ranking rules which might not be realistic, while the second approach is time consuming since it requires increasing the sample size from n to O(n 2 ). In this paper we consider an alternative approach that directly maintains a totally ordered set via projections. Our starting point is similar to that of Herbrich et. al [5] in the sense that we project each instance into the reals. However, our work then deviates and operates directly on rankings by associating each ranking with distinct sub-interval of the reals and adapting the support of each sub-interval while learning. In the next section we describe a simple and efficient online algorithm that manipulates concurrently the direction onto which we project the instances and the division into sub-intervals. In Sec. 3 we prove the correctness of the algorithm and analyze its performance in the mistake bound model. We describe in Sec. 4 experiments that compare the algorithm to online algorithms for classification and regression applied to ranking which demonstrate the merits of our approach. 2 The PRank Algorithm This paper focuses on online algorithms for ranking instances. We are given a sequence (Xl, yl), ... , (xt , yt) , ... of instance-rank pairs. Each instance xt is in IR n and its corresponding rank yt is an element from finite set y with a total order relation. We assume without loss of generality that y = {I , 2, ... ,k} with ">" as the order relation. The total order over the set Y induces a partial order over the instances in the following natural sense. We say that xt is preferred over X S if yt > yS. We also say that xt and x S are not comparable if neither yt > yS nor yt < yS. We denote this case simply as yt = yS. Note that the induced partial order is of a unique form in which the instances form k equivalence classes which are totally ordered l . A ranking rule H is a mapping from instances to ranks, H : IR n -+ y. The family of ranking rules we discuss in this paper employs a vector w E IR n and a set of k thresholds bl :::; ... :::; bk - l :::; bk = 00. For convenience we denote by b = (b l , . .. ,bk-d the vector of thresholds excluding bk which is fixed to 00. Given a new instance x the ranking rule first computes the inner-product between w and x . The predicted rank is then defined to be the index of the first (smallest) threshold br for which w . x < br . This type of ranking rules divide the space into parallel equally-ranked regions: all the instances that satisfy br - l < W? x < br are assigned the same rank r. Formally, given a ranking rule defined by wand b the predicted rank of an instance x is, H(x) = minrE{l, ... ,k}{r : w . x - br < O}. Note that the above minimum is always well defined since we set bk = 00. The analysis that we use in this paper is based on the mistake bound model for online learning. The algorithm we describe works in rounds. On round t the learning algorithm gets an instance xt. Given x t , the algorithm outputs a rank , il = minr {r : W? x - br < O}. It then receives the correct rank yt and updates its ranking rule by modifying wand b. We say that our algorithm made a ranking mistake if il f:. yt. IFor a discussion of this type of partial orders see [6] . Initialize: Set wI = 0 , b~ , ... , Loop: Fort=1 ,2, ... ,T ? ? ? ? bLl = 0, bl = 00 . Get a new rank-value xt E IRn. Predict fl = min r E{I, ... ,k} {r: w t . xt - b~ < o}. Get a new label yt. If fl t- yt update w t (otherwise set w t+! = w t , \;fr : b~+! = bn : 1. For r = 1, ... , k - 1 If yt :::; r Then y~ = -1 Else y~ = 1. 2. For r = 1, ... , k - 1 If (w t . xt - b~)y~ :::; 0 Then T; = y~ Else = o. 3. Update w t+! f- w t + CL r T;)xt. For r = 1, . .. , k - 1 update: b~+1 f- b~ - T; T; Output: H(x) = min r E{1, ... ,k} {r : w T +! . x - b;+! < O}. Figure 2: The PRank algorithm. We wish to make the predicted rank as close as possible to the true rank. Formally, the goal of the learning algorithm is to minimize the ranking-loss which is defined to be the number of thresholds between the true rank and the predicted rank. Using the representation of ranks as integers in {I ... k}, the ranking-loss after T rounds is equal to the accumulated difference between the predicted and true rank-values, '?'[=1 W- yt I. The algorithm we describe updates its ranking rule only on rounds on which it made ranking mistakes. Such algorithms are called conservative. We now describe the update rule of the algorithm which is motivated by the perceptron algorithm for classification and hence we call it the PRank algorithm (for Perceptron Ranking). For simplicity, we omit the index of the round when referring to an input instance-rank pair (x, y) and the ranking rule wand h. Since b1 :::; b2 :::; ... :::; bk - 1 :::; bk then the predicted rank is correct if w . x > br for r = 1, ... ,y - 1 and w . x < br for y, . .. , k - 1. We represent the above inequalities by expanding the rank y into into k - 1 virtual variables Yl , ... ,Yk-l. We set Yr = +1 for the case W? x > br and Yr = -1 for w . x < br . Put another way, a rank value y induces the vector (Yl, ... , Yk-d = (+1, ... , +1 , -1 , ... , -1) where the maximal index r for which Yr = +1 is y-1. Thus, the prediction of a ranking rule is correct if Yr(w? x - br ) > 0 for all r. If the algorithm makes a mistake by ranking x as fj instead of Y then there is at least one threshold, indexed r, for which the value of W? x is on the wrong side of br , i.e. Yr(w? x - br ) :::; O. To correct the mistake, we need to "move" the values of W? x and br toward each other. We do so by modifying only the values of the br's for which Yr (w . x - br ) :::; 0 and replace them with br - Yr. We also replace the value of w with w + ('? Yr)x where the sum is taken over the indices r for which there was a prediction error, i.e., Yr (w . x - br ) :::; o. An illustration of the update rule is given in Fig 1. In the example, we used the set Y = {I ... 5}. (Note that b5 = 00 is omitted from all the plots in Fig 1.) The correct rank of the instance is Y = 4, and thus the value of w . x should fall in the fourth interval, between b3 and b4 . However, in the illustration the value of w . x fell below b1 and the predicted rank is fj = 1. The threshold values b1 , b2 and b3 are a source of the error since the value of b1 , b2 , b3 is higher then W? x. To mend the mistake the algorithm decreases b1 , b2 and b3 by a unit value and replace them with b1 -1 , b2 -1 and b3 -1. It also modifies w to be w+3x since '?r:Yr(w.x- br):SOYr = 3. Thus, the inner-product W? x increases by 311x11 2 . This update is illustrated at the middle plot of Fig. 1. The updat ed prediction rule is sketched on the right hand side of Fig. 1. Note that after the update, the predicted rank of x is Y = 3 which is closer to the true rank y = 4. The pseudocode of algorithm is given in Fig 2. To conclude this section we like to note that PRank can be straightforwardly combined with Mercer kernels [8] and voting techniques [4] often used for improving the performance of margin classifiers in batch and online settings. 3 Analysis Before we prove the mistake bound of the algorithm we first show that it maintains a consistent hypothesis in the sense that it preserves the correct order of the thresholds. Specifically, we show by induction that for any ranking rule that can be derived by the algorithm along its run, (w 1 , b 1 ) , ... , (w T +1 , b T +1) we have that b~ :S ... :S bL1 for all t. Since the initialization of the thresholds is such that b~ :S b~ :S ... :S bL1' then it suffices to show that the claim holds inductively. For simplicity, we write the updating rule of PRank in an alternative form. Let [7f] be 1 if the predicate 7f holds and 0 otherwise. We now rewrite the value of (from Fig. 2) as = y~[(wt . xt - bny~ :S 0]. Note that the values of b~ are integers for all r and t since for all r we initialize b; = 0, and b~+l - b~ E {-1 , 0, +1}. T; T; Lemma 1 (Order Preservation) Let w t and b t be the current ranking rule, where bi :S .. . :S bL1' and let (xt,yt) be an instance-rank pair fed to PRank on round t. Denote by wt+1 and bt+1 the resulting ranking rule after the update of PRank, then bi+1 :S ... :S bt~ll? Proof: In order to show that PRank maintains the order of the thresholds we use the definition of the algorithm for y~, namely we define y~ = +1 for r < yt and y~ = -1 for r 2:: yt. We now prove that b~t~ 2:: b~+l for all r by showing that b~+l - b~ 2:: y~+1[(wt . xt - b~+1)Y;+l :S 0] - y;[(wt . xt - b;)y; :S 0], (1) which we obtain by substituting the values of bt+1. Since b~+1 :S b~ and b~ ,b~+1 E Z we get that the value of b~+1 - b~ on the left hand side of Eq. (1) is a non-negative integer. Recall that y~ = 1 if yt > r and y~ = -1 otherwise, and therefore, y~+l :S y~. We now analyze two cases. We first consider the case y~+1 :j:. y~ which implies that y~+l = -1, y~ = +1. In this case, the right hand-side of Eq. (1) is at most zero, and the claim trivially holds. The other case is when y~+1 = y~. Here we get that the value of the right hand-side Eq. (1) cannot exceed 1. We therefore have to consider only the case where b~ = b~+1 and y~+1 = y~. But given these two conditions we have that y~+1[(wt. xt - b~+1)Y~+1 < 0] and y~[(wt. xt - b~)y~ < 0] are equal. The right hand side of Eq. (1) is now zero and the inequality holds with ? equality. In order to simplify the analysis of the algorithm we introduce the following notation. Given a hyperplane wand a set of k -1 thresholds b we denote by v E ~n+k-1 the vector which is a concatenation of wand b that is v = (w, b). For brevity we refer to the vector vas a ranking rule. Given two vectors v' = (w', b ' ) and v = (w, b) we have v' . v = w' . w + b' . b and IIvl1 2 = IIwl1 2 + IlbW. Theorem 2 (Mistake bound) Let (xl, y1), ... , (x T , yT) be an input sequence for PRank where xt E ~n and yt E {l. .. k}. Denote by R2 = maxt Ilxtl12. Assume that there is a ranking rule v* = (w* , b*) with :S ... :S bk- 1 of a unit norm that classifies the entire sequence correctly with margin "( = minr,t{ (w* . xt - b;)yn > o. Then, the rank loss of the algorithm '?;=1 Iyt - yt I, is at most (k - 1) (R 2 + 1) / "(2 . br Proof: Let us fix an example (xt, yt) which the algorithm received on round t. By definition the algorithm ranked the example using the ranking rule v t which is composed of w t and the thresholds b t . Similarly, we denote by vt+l the updated rule (wt+l , bt+l) after round t. That is wt+l = w t + (" and bt+l = btr - Ttr , ur Tt)xt r r for r = 1, 2, ... , k - 1. Let us denote by n t = yt 1the difference between the true rank and the predicted rank. It is straightforward to verify that nt = 2:=r ITn Note that if there wasn't a ranking mistake on round t then = for r = 1, ... , k-1, and thus also nt = 0. To prove the theorem we bound 2:=t nt from above by bounding IIvtl12 from above and below. First, we derive a lower bound on IIvtl12 by bounding v* . v H1 . Substituting the values of w H1 and b H1 we get, W- T; ? k-1 v* . vt+l = v* . v t + 2:= T; (w* . xt - b;) (2) r=1 T; from the pseudocode in Fig. 2 we need to analyze two cases. If (w t ?xt - b~)y; :::; ? then T; = y;. Using the assumption that v* ranks the data correctly with a margin We further bound the right term by considering two cases. Using the definition of of at least "( we get that T;(W* . xt - b;) ~ "(. For the other case for which (w t . xt - b;)y; > we have T; = and thus T;(W* . xt - b;) = 0. Summing now over r we get, ? ? k-1 2:= T; (w* . x t - b;) ~ nt"( . (3) r= 1 Combining Eq. (2) and Eq. (3) we get v* . vt+l ~ v* . v t + nt"(. Unfolding the sum, we get that after T rounds the algorithm satisfies, v* . v T+ 1 ~ 2:=t nt"( = "( 2:=t nt. Plugging this result into Cauchy-Schwartz inequality, (1Iv T+1 11 21Iv* 112 ~ (vT+l . v*) 2) and using the assumption that v* is of a unit norm we get the lower bound, IIv T+ll1 2 ~ (2:=t nt)2 "(2. Next, we bound the norm of v from above. As before, assume that an example (xt, yt) was ranked using the ranking rule v t and denote by vt+l the ranking rule after the round. We now expand the values ofw t+1 and bt+l in the norm ofv H1 and get, IIv H1 112 = IIwtl12 + IIb t l1 2 + 2 2:=r T; (w t . xt - b;) + (2:=r T;)21IxtI12 + 2:=r (T;)2. Since T; E {-1,0,+1} we have that (2:=rT;)2 :::; (nt)2 and 2:=r(T;) 2 = nt and we therefore get, IIv H1 112 :::; IIvtl12 + 22:= T; (w t . xt - b~) + (nt)21IxtW + nt . (4) r We further develop the second term using the update rule of the algorithm and get, 2:= T; (w t . xt - b~) = 2:=[(wt . xt - b~)y; :::; 0] ((wt . xt - b~)y~) :::; r ?. (5) r Plugging Eq. (5) into Eq. (4) and using the bound IIxtl12 :::; R2 we get that IlvH1112:::; IIvtl12 + (nt)2R2 + nt. Thus, the ranking rule we obtain after T rounds of the algorithm satisfies the upper bound, IlvT+l W :::; R2 2:=t(nt )2 + 2:=t nt. Combining the lower bound IlvT+l W ~ (2:=t nt)2 "(2 with the upper bound we have that, (2:=tnt) 2"(2:::; Ilv T+1112:::; R2 2:=t(nt )2 + 2:=t nt . Dividing both sides by "(2 2:=tnt we finally get, 2:= nt :::; R2 [2:=t(n t )2] t f [2:=t ntl + 1 . (6) "( By definition, nt is at most k - 1, which implies that 2:=t(n t )2 :::; 2:=t nt(k - 1) = (k -1) 2:=t nt. Using this inequality in Eq. (6) we get the desired bound, 2:=;=1 Ig t ytl = 2:=;=1 nt :::; [(k - 1)R2 + 1lh2 :::; [(k - 1)(R2 + 1)lh2 . ? i" I ... ~ Figure 3: Comparison of the time-averaged ranking-loss of PRank, WH, and MCP on synthetic data (left). Comparison of the time-averaged ranking-loss of PRank, WH, and MCP on the EachMovie dataset using viewers who rated and at least 200 movies (middle) and at least 100 movies (right). 4 Experiments In this section we describe experiments we performed that compared PRank with two other online learning algorithms applied to ranking: a multiclass generalization of the perceptron algorithm [2], denoted MCP, and the Widrow-Hoff [9] algorithm for online regression learning which we denote by WHo For WH we fixed its learning rate to a constant value. The hypotheses the three algorithms maintain share similarities but are different in their complexity: PRank maintains a vector w of dimension n and a vector of k - 1 modifiable thresholds b, totaling n + k - 1 parameters; MCP maintains k prototypes which are vectors of size n, yielding kn parameters; WH maintains a single vector w of size n. Therefore, MCP builds the most complex hypothesis of the three while WH builds the simplest. Due to the lack of space, we only describe two sets of experiments with two different datasets. The dataset used in the first experiment is synthetic and was generated in a similar way to the dataset used by Herbrich et. al. [5]. We first generated random points x = (Xl, X2) uniformly at random from the unit square [0,1 f. Each point was assigned a rank y from the set {I, ... , 5} according to the following ranking rule, y = maxr{r : lO((XI - 0.5)(X2 - 0.5)) + ~ > br } where b = (-00 , -1, -0.1,0.25,1) and ~ is a normally distributed noise of a zero mean and a standard deviation of 0.125. We generated 100 sequences of instance-rank pairs each of length 7000. We fed the sequences to the three algorithms and obtained a prediction for each instance. We converted the real-valued predictions of WH into ranks by rounding each prediction to its closest rank value. As in ~5] we used a non-homogeneous polynomial of degree 2, K(XI' X2) = ((Xl? X2) + 1) as the inner-product operation between each input instance and the hyperplanes the three algorithms maintain. At each time step, we computed for each algorithm the accumulated ranking-loss normalized by the instantaneous sequence length. Formally, the time-averaged loss Iyt _ytl. We computed these losses for T = 1, ... ,7000. after T rounds is, (liT) To increase the statistical significance of the results we repeated the process 100 times, picking a new random instance-rank sequence of length 7,000 each time, and averaging the instantaneous losses across the 100 runs. The results are depicted on the left hand side of Fig. 3. The 95% confidence intervals are smaller then the symbols used in the plot. In this experiment the performance of MPC is constantly worse than the performance of WH and PRank. WH initially suffers the smallest instantaneous loss but after about 500 rounds PRank achieves the best performance and eventually the number of ranking mistakes that PRank suffers is significantly lower than both WH and MPC. 'L,i In the second set of experiments we used the EachMovie dataset [7]. This dataset is used for collaborative filtering tasks and contains ratings of movies provided by 61 , 265 people. Each person in the dataset viewed a subset of movies from a collection of 1623 titles. Each viewer rated each movie that she saw using one of 6 possible ratings: 0, 0.2, 0.4, 0.6, 0.8,1. We chose subsets of people who viewed a significant amount of movies extracting for evaluation people who have rated at least 100 movies. There were 7,542 such viewers. We chose at random one person among these viewers and set the person's ratings to be the target rank. We used the ratings of all the rest of the people who viewed enough movies as features. Thus, the goal is to learn to predict the "taste" of a random user using the user's past ratings as a feedback and the ratings of fellow viewers as features. The prediction rule associates a weight with each fellow viewer an therefore can be seen as learning correlations between the tastes of different viewers. Next, we subtracted 0.5 from each rating and therefore the possible ratings are -0.5 , -0.3, -0.1 , 0.1, 0.3, 0.5. This linear transformation enabled us to assign a value of zero to movies which have not been rated. We fed these feature-rank pairs one at a time, in an online fashion . Since we picked viewer who rated at least 100 movies, we were able to perform at least 100 rounds of online predictions and updates. We repeated this experiment 500 times, choosing each time a random viewer for the target rank. The results are shown on the right hand-side of Fig. 3. The error bars in the plot indicate 95% condfidence levels. We repeated the experiment using viewers who have seen at least 200 movies. (There were 1802 such viewers.) The results of this experiment are shown in the middle plot of Fig. 3. Along the entire run of the algorithms , PRank is significantly better than WH, and consistently better than the multiclass perceptron algorithm, although the latter employs a bigger hypothesis. Finally, we have also evaluated the performance of PRank in a batch setting, using the experimental setup of [5]. In this experiment, we ran PRank over the training data as an online algorithm and used its last hypothesis to rank unseen test data. Here as well PRank came out first, outperforming all the algorithms described in [5]. Acknowledgments Thanks to Sanjoy Dagupta and Rob Schapire for numerous discussions on ranking problems and algorithms. Thanks also to Eleazar Eskin and Uri Maoz for carefully reading the manuscript. References [1] William W. Cohen, Robert E. Schapire, and Yoram Singer. Learning to order things. Journal of Artificial Int elligence Research, 10:243- 270 , 1999. [2] K. Crammer and Y. Singer. Ultraconservative online algorithms for multiclass problems. Proc. of the Fourteenth Annual ConI on Computational Learning Theory, 200l. [3] Y. Freund, R. Iyer, R. E. Schapire, and Y. Singer. An efficient boosting algorithm for combining preferences. Machine Learning: Proc. of the Fifteenth Inti. ConI, 1998. [4] Y. Freund and R. E. Schapire. Large margin classification using the perceptron algorithm. Machine Learning, 37(3): 277-296, 1999. [5] R. Herbrich, T. Graepel, and K. Obermayer. Large margin rank boundaries for ordinal regression. Advances in Large Margin Classifiers. MIT Press, 2000. [6] J. Kemeny and J . Snell. Mathematical Models in the Social Sciences. MIT Press, 1962. [7] Paul McJones. EachMovie collaborative filtering data set. DEC Systems Research Center, 1997. http://www.research.digital.com/SRC/eachmoviej. [8] Vladimir N. Vapnik. Statistical Learning Theory. Wiley, 1998. [9] Bernard Widrow and Marcian E. Hoff. Adaptive switching circuits. 1960 IRE WESCON Convention Record, 1960. Reprinted in Neurocomputing (MIT Press, 1988).
2023 |@word middle:3 polynomial:1 norm:4 bn:1 carry:1 contains:1 outperforms:1 past:2 current:1 com:1 nt:24 must:1 realistic:1 plot:5 update:13 yr:10 item:2 short:1 record:1 eskin:1 ire:1 boosting:1 herbrich:3 hyperplanes:1 preference:1 mathematical:2 along:2 prove:5 introduce:1 nor:1 considering:1 totally:3 increasing:1 project:2 classifies:1 notation:1 provided:1 circuit:1 israel:1 what:1 transformation:1 fellow:2 voting:1 wrong:1 classifier:2 schwartz:1 unit:4 normally:1 omit:1 yn:1 before:2 eleazar:1 engineering:1 mistake:12 switching:1 might:2 chose:2 therein:1 initialization:1 equivalence:1 bi:2 averaged:3 unique:1 acknowledgment:1 adapting:1 significantly:2 projection:1 confidence:1 get:17 cannot:2 close:2 onto:1 convenience:1 put:1 www:1 yt:22 center:1 modifies:1 jerusalem:1 economics:1 starting:1 straightforward:1 simplicity:2 assigns:1 manipulates:1 rule:29 enabled:1 updated:1 target:2 user:4 homogeneous:1 hypothesis:5 associate:1 element:1 iib:1 updating:1 mend:1 region:1 ordering:1 decrease:1 yk:2 ran:1 src:1 complexity:2 inductively:1 iiv:3 rewrite:1 division:1 ofv:1 distinct:1 describe:9 artificial:1 choosing:1 whose:1 pref:1 valued:1 say:3 otherwise:3 unseen:1 btr:1 online:13 sequence:7 product:3 maximal:1 fr:1 loop:1 combining:3 maoz:1 object:1 ilvt:2 derive:1 develop:1 ac:1 widrow:2 school:1 received:1 eq:9 dividing:1 c:1 predicted:9 implies:2 indicate:1 convention:1 direction:1 correct:7 modifying:2 virtual:1 assign:2 suffices:1 fix:1 generalization:1 snell:1 viewer:11 hold:4 mapping:1 predict:3 claim:2 substituting:2 achieves:1 smallest:2 omitted:1 ilv:1 proc:2 label:5 title:1 saw:1 correctness:2 unfolding:1 mit:3 clearly:1 concurrently:1 always:1 rather:1 totaling:1 ytl:2 derived:1 focus:1 she:1 consistently:1 rank:46 underscore:1 sense:3 accumulated:2 bt:6 entire:2 initially:1 relation:3 irn:1 expand:1 semantics:1 sketched:1 x11:1 classification:7 among:1 denoted:1 initialize:2 hoff:2 field:1 equal:2 lit:1 koby:1 simplify:1 inherent:1 employ:2 bl1:3 composed:1 preserve:1 neurocomputing:1 maintain:2 william:1 interest:1 evaluation:1 yielding:1 closer:1 partial:3 indexed:1 iv:2 divide:1 desired:1 instance:25 classify:1 cover:1 minr:2 deviation:1 subset:2 predicate:1 rounding:1 straightforwardly:1 kn:1 synthetic:3 combined:1 thoroughly:1 referring:1 person:3 thanks:2 huji:1 pranking:1 yl:3 picking:1 lh2:2 worse:1 book:1 sidestep:1 converted:1 sec:2 b2:5 int:1 satisfy:1 ranking:46 performed:2 view:2 h1:6 picked:1 analyze:4 maintains:6 parallel:1 collaborative:5 minimize:1 il:3 ir:3 square:1 who:7 suffers:2 ed:1 definition:4 mpc:2 associated:1 proof:2 dataset:7 wh:10 recall:1 graepel:1 carefully:1 manuscript:1 higher:1 evaluated:1 generality:2 erences:1 correlation:1 hand:7 receives:1 lack:2 b3:5 verify:1 true:6 normalized:1 hence:1 assigned:2 equality:1 illustrated:1 round:15 ll:1 ttr:1 theoretic:2 demonstrate:1 tt:1 l1:1 fj:2 instantaneous:3 common:2 pseudocode:2 overview:1 cohen:1 b4:1 refer:2 significant:1 trivially:1 similarly:1 kobics:1 similarity:1 closest:1 inequality:4 outperforming:1 came:1 vt:5 seen:2 minimum:1 determine:1 itn:1 preservation:1 bother:1 eachmovie:4 retrieval:1 equally:1 y:4 bigger:1 plugging:2 va:1 prediction:9 regression:7 metric:1 fifteenth:1 kernel:1 represent:1 dec:1 interval:6 else:2 source:2 rest:1 fell:1 induced:1 thing:1 integer:4 call:1 extracting:1 exceed:1 enough:1 associating:1 reduce:1 inner:3 prototype:1 wasn:1 br:21 multiclass:3 reprinted:1 whether:1 motivated:1 b5:1 constitute:1 amount:1 ifor:1 induces:2 simplest:1 schapire:4 http:1 correctly:2 modifiable:1 write:1 threshold:12 neither:1 sum:2 wand:5 run:4 fourteenth:1 you:1 fourth:1 family:1 comparable:1 bound:15 fl:2 fan:1 annual:1 x2:4 min:2 structured:1 according:1 across:1 smaller:1 ur:1 wi:1 rob:1 inti:1 taken:1 mcp:5 ntl:1 discus:3 eventually:1 singer:5 ordinal:1 merit:1 fed:3 operation:1 subtracted:1 alternative:2 batch:2 running:1 coni:2 yoram:2 build:2 bl:2 move:1 rt:1 obermayer:1 kemeny:1 concatenation:1 cauchy:1 toward:1 induction:1 length:3 index:4 elligence:1 illustration:3 iyt:2 hebrew:1 vladimir:1 setup:1 robert:1 negative:1 ofw:1 perform:1 upper:2 datasets:1 finite:1 prank:20 excluding:1 bll:1 y1:1 tnt:2 rating:14 bk:8 cast:1 pair:6 fort:1 namely:1 able:1 suggested:1 bar:1 below:2 reading:1 natural:2 ranked:3 ll1:1 movie:14 rated:5 numerous:1 deviate:1 taste:2 freund:2 loss:12 filtering:5 iiwl1:1 digital:1 degree:2 consistent:1 imposes:1 mercer:1 intractability:1 share:2 maxt:1 lo:1 last:1 side:9 perceptron:5 fall:1 distributed:1 feedback:1 dimension:1 boundary:1 computes:1 made:2 collection:1 adaptive:1 ig:1 social:2 preferred:1 maxr:1 wescon:1 b1:6 summing:1 conclude:1 consuming:1 xi:2 ultraconservative:1 learn:1 expanding:1 improving:1 cl:1 complex:1 significance:1 main:1 bounding:2 noise:1 paul:1 repeated:3 fig:10 fashion:1 wiley:1 sub:3 wish:1 xl:4 theorem:2 xt:30 showing:1 symbol:1 r2:8 vapnik:1 iyer:1 uri:1 margin:6 gap:1 updat:1 depicted:1 simply:1 expressed:1 ordered:3 satisfies:2 constantly:1 goal:6 viewed:3 replace:3 specifically:1 operates:1 uniformly:1 wt:10 hyperplane:1 averaging:1 lemma:1 conservative:1 total:4 called:1 sanjoy:1 bernard:1 experimental:1 meaningful:1 formally:3 support:1 people:4 latter:2 crammer:2 brevity:1 ongoing:1
1,123
2,024
Incorporating Invariances in Nonlinear Support Vector Machines Olivier Chapelle Bernhard Scholkopf olivier.chapelle@lip6.fr LIP6, Paris, France Biowulf Technologies bernhard.schoelkopf@tuebingen.mpg.de Max-Planck-Institute, Tiibingen, Germany Biowulf Technologies Abstract The choice of an SVM kernel corresponds to the choice of a representation of the data in a feature space and, to improve performance , it should therefore incorporate prior knowledge such as known transformation invariances. We propose a technique which extends earlier work and aims at incorporating invariances in nonlinear kernels. We show on a digit recognition task that the proposed approach is superior to the Virtual Support Vector method, which previously had been the method of choice. 1 Introduction In some classification tasks, an a priori knowledge is known about the invariances related to the task. For instance, in image classification, we know that the label of a given image should not change after a small translation or rotation. More generally, we assume we know a local transformation Lt depending on a parameter t (for instance, a vertical translation of t pixels) such that any point x should be considered equivalent to LtX, the transformed point. Ideally, the output of the learned function should be constant when its inputs are transformed by the desired invariance. It has been shown [1] that one can not find a non-trivial kernel which is globally invariant. For this reason, we consider here local invariances and for this purpose we associate at each training point X i a tangent vector dXi, dXi = lim -1 (LtXi t--+o t Xi) 81t=o LtXi = - 8t In practice dXi can be either computed by finite difference or by differentiation. Note that generally one can consider more than one invariance transformation. A common way of introducing invariances in a learning system is to add the perturbed examples LtXi in the training set [7]. Those points are often called virtual examples. In the SVM framework , when applied only to the SVs, it leads to the Virtual Support Vector (VSV) method [10]. An alternative to this is to modify directly the cost function in order to take into account the tangent vectors. This has been successfully applied to neural networks [13] and linear Support Vector Machines [11]. The aim of the present work is to extend these methods to the case of nonlinear SVMs which will be achieved mainly by using the kernel peA trick [12]. The paper is organized as follows. After introducing the basics of Support Vector Machines in section 2, we recall the method proposed in [11] to train invariant linear SVMs (section 3). In section 4, we show how to extend it to the nonlinear case and finally experimental results are provided in section 5. 2 Support Vector Learning We introduce some standard notations for SVMs; for a complete description, see [15]. Let {(Xi, Yi) h<i<n be a set of training examples, Xi E IRd , belonging to classes labeled by Yi E {-1,1}. In kernel methods, we map these vectors into a feature space using a kernel function K(Xi' Xj) that defines an inner product in this feature space. The decision function given by an SVM is the maximal margin hyperplane in this space, g(X) = sign(f(x)), where f(x) = (~a?YiK(Xi'X) + b) . (1) The coefficients a? are obtained by maximizing the functional 1 n W(o:) = Lai -"2 i=l n L aiajYiyjK(Xi,Xj) i,j=l under the constraints L:~= 1 aiYi = 0 and ai (2) ~ O. This formulation of the SVM optimization problem is called the hard margin formulation since no training errors are allowed. In the rest of the paper, we will not consider the so called soft-margin SVM algorithm [4], where training errors are allowed. 3 Invariances for Linear SVMs For linear SVMs, one wants to find a hyperplane whose normal vector w is as orthogonal as possible to the tangent vectors. This can be easily understood from the equality f(Xi + dXi) - f(Xi) = w . dXi' For this purpose, it has been suggested [11] to minimize the functional n (1 - ')')w 2 + ')' L(w, dXi)2 (3) i=l subject to the constraints Yi(W . Xi + b) ~ 1. The parameter,), trades off between normal SVM training (')' = 0) and full enforcement of the orthogonality between the hyperplane and the invariance directions (')' ---+ 1). Let us introduce c, ~ ((1-0)[ +0 ~dx'dxi) 'i', the square root of the regularized covariance matrix of the tangent vectors. (4) It was shown in [11] that training a linear invariant SVM, i.e. minimizing (3), is equivalent to a standard SVM training after the following linear transformation of the input space -1 X --+ C, x. This method led to significant improvements in linear SVMs, and to small improvements when used as a linear preprocessing step in nonlinear SVMs. The latter, however, was a hybrid system with unclear theoretical foundations. In the next section we show how to deal with the nonlinear case in a principled way. 4 Extension to the nonlinear case In the nonlinear case, the data are first mapped into a high-dimensional feature space where a linear decision boundary is computed. To extend directly the previous analysis to the nonlinear case, one would need to compute the matrix C, in feature space, C, = ( (1 - '"Y)I + '"Y ~ dlJ> (Xi) dlJ> (Xi) T and the new kernel function K(x , y) = C~ llJ>(x) . C~ llJ>(y) ) 1~ (5) = lJ>(x) T C~ 21J>(y) (6) However, due to the high dimension of the feature space, it is impossible to do it directly. We propose two different ways for overcoming this difficulty. 4.1 Decomposition of the tangent Gram matrix In order to be able to compute the new kernel (6) , we propose to diagonalize the matrix C, (eq 5) using a similar approach as the kernel PCA trick [12]. In that article, they showed how it was possible to diagonalize the feature space covariance matrix by computing the eigendecomposition of the Gram matrix of those points. Presently, instead of having a set of training points {1J>(Xi)} , we have a set of tangent vectors {dlJ> (Xi)} and a tangent covariance matrix (the right term of the sum in (5)) Let us introduce the Gram matrix Kt of the tangent vectors: Kij = dlJ>(Xi )? dlJ>(xj) K(Xi +dXi, Xj +dxj) - K(Xi +dXi, Xj) - K(Xi ' Xj +dxj) + K(Xi' Xj) (7) (8) . dxiT02K(Xi,Xj)d ~ ~ XJ UXiUXj This matrix Kt can be computed either by finite differences (equation 7) or with the analytical derivative expression given by equation (8) . Note that for a linear kernel, K(x,y) = x T y, and (8) reads Kfj = dxi dXj, which is a standard dot product between the tangent vectors. Writing the eigendecomposition of Kt as Kt = U AUT , and using the kernel PCA tools [12], one can show after some algebra (details in [2]) that the new kernel matrix reads K(x,y) 1 Kx y -I - '"Y (,) + n1( ~ -Ap U.dx,TOK(Xi' X)) (~ ~1 U~ ~ 'p ~ 1 '"Y Ap +1- (~U. ~ 'p ~1 '"Y - -1 -) 1 - '"Y T d OK(Xi' x, ~ U~ y)) 4.2 The kernel PCA map A drawback of the previous approach appears when one wants to deal with multiple invariances (i.e. more than one tangent vector per training point). Indeed, it requires to diagonalize the matrix Kt (cf eq 7), whose size is equal to the number of different tangent vectors. For this reason, we propose an alternative method. The idea is to use directly the so-called kernel peA map, first introduced in [12] and extended in [14]. This map is based on the fact that even in a high dimensional feature space 1i, a training set {Xl , .. . , x n } of size n when mapped to this feature space spans a subspace E C 1i whose dimension is at most n . More precisely, if (VI"'" Vn ) E En is an orthonormal basis of E with each Vi being a principal axis of {<I>(xd, ... , <I> (x n )} , the kernel peA map 'i/J : X -+ ~n is defined coordinatewise as 'i/Jp (x) = <I>(x) . v P ' 1:S p:S n. Each principal direction has a linear expansion on the training points {<I>(Xi)} and the coefficients of this expansion are obtained using kernel peA [12]. Writing the eigendecompostion of K as K = U AUT, with U an orthonormal matrix and A a diagonal one, it turns out that the the kernel peA map reads 'i/J(x) = A-1/2U T k(x), (9) where k (x) = (K(x, Xl)"'" K(x, xn)) T . Note that by definition , for all i and j , <I>(Xi) and <I>(Xj) lie in E and thus K(Xi ' Xj) = <I>(Xi) . <I>(Xj) = 'i/J(Xi) . 'i/J(Xj). This reflects the fact that if we retain all principal components, kernel peA is just a basis transform in E, leaving the dot product of training points invariant. As a consequence, training a nonlinear SVM on {Xl , ... , xn} is equivalent to training a linear SVM on {'i/J(xd, . . . ,'i/J(xn )} and thus, thanks to the nonlinear mapping 'i/J, we can work directly in the linear space E and use exactly the technique described for invariant linear SVMs (section 3) . However the invariance directions d<I>(Xi) do not necessarily belong to E. By projecting them onto E, some information might be lost. The hope is that this approximation will give a similar decision function to the exact one obtained in section 4.l. Finally, the proposed algorithm consists in training an invariant linear SVM as described in section 3 with training set { 'i/J(XI) , ... ,'i/J(xn)} and with invariance directions {d'i/J(XI) , ... , d'i/J (x n)}, where d'i/J (Xi) = 'i/J (Xi + dXi ) - 'i/J(Xi), which can be expressed from equation (9) as 4.3 Comparisons with the VSV method One might wonder what is the difference between enforcing an invariance and just adding the virtual examples LtXi in the training set. Indeed the two approaches are related and some equivalence can be shown [6] . So why not just add virtual examples? This is the idea of the Virtual Support Vector (VSV) method [10] . The reason is the following: if a training point Xi is far from the margin, adding the virtual example LtXi will not change the decision boundary since neither of the points can become a support vector. Hence adding virtual examples in the SVM framework enforces invariance only around the decision boundary, which, as an aside, is the main reason why the virtual SV method only adds virtual examples generated from points that were support vectors in the earlier iteration. One might argue that the points which are far from the decision boundary do not provide any information anyway. On the other hand, there is some merit in not only keeping the output label invariant under the transformation Lt, but also the real-valued output. This can be justified by seeing the distance of a given point to the margin as an indication of its class-conditional probability [8]. It appears reasonable that an invariance transformation should not affect this probability too much. 5 Experiments In our experiments, we compared a standard SVM with several methods taking into account invariances: standard SVM with virtual examples (cf. the VSV method [10]) [VSV], invariant SVM as described in section 4.1 [ISVM] and invariant hyperplane in kernel peA coordinates as described in section 4.2 [ IHKPcA ]. The hybrid method described in [11] (see end of section 3) did not perform better than the VSV method and is not included in our experiments for this reason. Note that in the following experiments, each tangent vector d<I>(Xi) has been normalized by the average length Ild<I>(xi)W/n in order to be scale independent. vI: 5.1 Toy problem The toy problem we considered is the following: the training data has been generated uniformly from [-1 , 1]2. The true decision boundary is a circle centered at the origin: f(x) = sign(x2 - 0.7). The a priori knowledge we want to encode in this toy problem is local invariance under rotations. Therefore, the output of the decision function on a given training point Xi and on its image R(Xi,C:) obtained by a small rotation should be as similar as possible. To each training point, we associate a tangent vector dXi which is actually orthogonal to Xi. A training set of 30 points was generated and the experiments were repeated 100 times. A Gaussian kernel K(x,y) = exp (_ II X2~~ 1I 2) was chosen. The results are summarized in figure 1. Adding virtual examples (VSV method) is already very useful since it made the test error decrease from 6.25% to 3.87% (with the best choice of a). But the use of ISVM or IHKPcA yields even better performance. On this toy problem, the more the invariances are enforced b -+ 1), the better the performances are (see right side of figure 1), reaching a test error of 1.11%. When comparing log a = 1.4 and log a = 0 (right side of of figure 1), one notices that the decrease in the test error does not have the same speed. This is actually the dual of the phenomenon observed on the left side of this figure: for a same value of gamma, the test error tends to increase, when a is larger. This analysis suggests that 'Y needs to be adapted in function of a. This can be done automatically by the gradient descent technique described in [3]. 0.12 - 0. 14 - - log sigma=-O.8 log sigma=O 10 si ma= 1,4 0 .12 0.1 0.06 0 .08 0.04 0.02 O.02 '----_~ , ------: -0~ .5--~ 0 --0~ .5,------~------'c-" .5 Log sigma %'------~-~-~ 6 -~ 8--,~ 0 -~ , 2~ - Log (1-gamma) Figure 1: Left: test error for different learning algorithms plotted against the width of a RBF kernel and "( fixed to 0.9. Right: test error of IHKPcA across "( and for different values of (5. The test errors are averaged over the 100 splits and the error bars correspond to the standard deviation of the means. 5.2 Handwritten digit recognition As a real world experiment, we tried to incorporate invariances for a handwritten digit recognition task. The USPS dataset have been used extensively in the past for this purpose, especially in the SVM community. It consists of 7291 training and 2007 test examples. According to [9], the best performance has been obtained for a polynomial kernel of degree 3, and all the results described in this section were performed using this kernel. The local transformations we considered are translations (horizontal and vertical). All the tangent vectors have been computed by a finite difference between the original digit and its I-pixel translated. We split the training set into 23 subsets of 317 training examples after a random permutation of the training and test set. Also we concentrated on a binary classification problem, namely separating digits a to 4 against 5 to 9. The gain in performance should also be valid for the multiclass case. Figure 2 compares ISVM, IHKPcA and VSV for different values of "(. From those figures, it can be seen that the difference between ISVM (the original method) and IHKPcA (the approximation) is much larger than in the toy example. The difference to the toy example is probably due to the input dimensionality. In 2 dimensions, with an RBF kernel, the 30 examples of the toy problem "almost span" the whole feature space, whereas with 256 dimensions , this is no longer the case. What is noteworthy in these experiments is that our proposed method is much better than the standard VSV. As explained in section 4.3, the reason for this might be that invariance is enforced around all training points and not only around support vectors. Note that what we call VSV here is a standard SVM with a double size training set containing the original data points and their translates. The horizontal invariance yields larger improvements than the vertical one. One of the reason might be that the digits in the USPS database are already centered vertically. - 0 .068 - - IHKPCA ISVM VSV - - 0 .066 0.066 0 .064 0 .064 0 .062 0.062 0.06 0.06 0 .058 0 .058 0.056 0.056 0 .054 0 0.5 1.5 2 2.5 3.5 -Log (1-gamma) Vertical translation (to the top) - 0 .068 0 .054 0 0.5 1.5 2 2.5 IHKPCA ISVM VSV 3.5 -Log (1-gamma) Horizontal translation (to the right) Figure 2: Comparison of ISVM, IHKPcA and VSV on the USPS dataset. The left of the plot ("( = 0) corresponds to standard SVM whereas the right part of the plot h -+ 1) means that a lot of emphasis is put on the enforcement of the constraints. The test errors are averaged over the 23 splits and the error bars correspond to the standard deviation of the means. 6 Conclusion We have extended a method for constructing invariant hyperplanes to the nonlinear case. We have shown results that are superior to the virtual SV method. The latter has recently broken the record on the NIST database which is the "gold standard" of handwritten digit benchmarks [5], therefore it appears promising to also try the new system on that task. For this propose, a large scale version of this method needs to be derived. The first idea we tried is to compute the kernel PCA map using only a subset of the training points. Encouraging results have been obtained on the lO-class USPS database (with the whole training set), but other methods are also currently under study. References [1] C. J. C. Burges. Geometry and invariance in kernel based methods. In B. Sch6lkopf, C. J . C. Burges, and A. J . Smola, editors, Advances in Kernel Methods - Support Vector Learning. MIT Press, 1999. [2] O. Chapelle and B. Sch6lkopf. Incorporating invariances in nonlinear Support Vector Machines, 2001. Availabe at: www-connex.lip6.frrchapelle. [3] O. Chapelle, V. Vapnik, O. Bousquet, and S. Mukherjee. Choosing multiple parameters for support vector machines. Machine Learning, 46:131- 159, 2002. [4] C. Cortes and V. Vapnik. Support vector networks. Machine Learning, 20:273 297,1995. [5] D. DeCoste and B. Sch6lkopf. Training invariant support vector machines. Machine Learning, 2001. In press. [6] Todd K. Leen. From data distributions to regularization in invariant learning. In Nips, volume 7. The MIT Press, 1995. [7] P. Niyogi, T. Poggio, and F. Girosi. Incorporating prior information in machine learning by creating virtual examples. IEEE Proceedings on Intelligent Signal Processing, 86(11):2196-2209, November 1998. [8] John Platt. Probabilities for support vector machines. In A. Smola, P. Bartlett, B. Sch6lkopf, and D. Schuurmans, editors, Advances in Large Margin Classifiers. MIT Press, Cambridge, MA, 2000. [9] B. Sch6lkopf, C. Burges, and V. Vapnik. Extracting support data for a given task. In U. M. Fayyad and R. Uthurusamy, editors, First International Conference on Knowledge Discovery fj Data Mining. AAAI Press, 1995. [10] B. Sch6lkopf, C. Burges, and V. Vapnik. Incorporating invariances in support vector learning machines. In Artificial Neural Networks - ICANN'96, volume 1112, pages 47- 52, Berlin, 1996. Springer Lecture Notes in Computer Science. [11] B. Sch6lkopf, P. Y. Simard, A. J. Smola, and V. N. Vapnik. Prior knowledge in support vector kernels. In MIT Press, editor, NIPS, volume 10, 1998. [12] B. Sch6lkopf, A. Smola, and K.-R. Muller. Nonlinear component analysis as a kernel eigenvalue problem. Neural Computation, 10:1299- 1310, 1998. [13] P. Simard, Y. LeCun, J. Denker, and B. Victorri. Transformation invariance in pattern recognition, tangent distance and tangent propagation. In G. Orr and K. Muller, editors, Neural Networks: Tricks of the trade. Springer, 1998. [14] K. Tsuda. Support vector classifier with asymmetric kernel function. In M. Verleysen, editor, Proceedings of ESANN'99, pages 183- 188,1999. [15] V. Vapnik. Statistical Learning Theory. John Wiley & Sons, 1998.
2024 |@word version:1 polynomial:1 tried:2 covariance:3 decomposition:1 past:1 comparing:1 si:1 dx:2 john:2 girosi:1 plot:2 aside:1 record:1 hyperplanes:1 become:1 scholkopf:1 consists:2 introduce:3 indeed:2 mpg:1 globally:1 automatically:1 encouraging:1 decoste:1 provided:1 notation:1 what:3 transformation:8 differentiation:1 xd:2 exactly:1 classifier:2 platt:1 planck:1 understood:1 local:4 modify:1 tends:1 vertically:1 consequence:1 todd:1 ap:2 noteworthy:1 might:5 emphasis:1 equivalence:1 suggests:1 averaged:2 lecun:1 enforces:1 practice:1 lost:1 digit:7 seeing:1 onto:1 put:1 impossible:1 writing:2 www:1 equivalent:3 map:7 maximizing:1 orthonormal:2 anyway:1 coordinate:1 exact:1 olivier:2 origin:1 associate:2 trick:3 recognition:4 asymmetric:1 mukherjee:1 labeled:1 database:3 observed:1 schoelkopf:1 trade:2 decrease:2 principled:1 broken:1 ideally:1 algebra:1 basis:2 usps:4 translated:1 easily:1 train:1 isvm:7 artificial:1 choosing:1 whose:3 larger:3 valued:1 niyogi:1 transform:1 indication:1 eigenvalue:1 analytical:1 propose:5 product:3 maximal:1 fr:1 gold:1 description:1 double:1 depending:1 connex:1 eq:2 esann:1 direction:4 drawback:1 pea:7 centered:2 virtual:14 extension:1 around:3 considered:3 normal:2 exp:1 mapping:1 purpose:3 label:2 currently:1 lip6:3 successfully:1 tool:1 reflects:1 hope:1 mit:4 gaussian:1 aim:2 reaching:1 encode:1 derived:1 improvement:3 mainly:1 lj:1 transformed:2 france:1 germany:1 pixel:2 classification:3 dual:1 priori:2 verleysen:1 equal:1 having:1 intelligent:1 gamma:4 geometry:1 n1:1 mining:1 kt:5 poggio:1 orthogonal:2 desired:1 circle:1 plotted:1 tsuda:1 theoretical:1 instance:2 kij:1 earlier:2 soft:1 cost:1 introducing:2 deviation:2 subset:2 wonder:1 too:1 perturbed:1 sv:2 thanks:1 international:1 retain:1 off:1 aaai:1 containing:1 creating:1 derivative:1 simard:2 toy:7 account:2 de:1 orr:1 summarized:1 coefficient:2 vi:3 performed:1 root:1 lot:1 try:1 minimize:1 square:1 yield:2 correspond:2 handwritten:3 definition:1 against:2 dxi:12 gain:1 dataset:2 recall:1 knowledge:5 lim:1 dimensionality:1 organized:1 actually:2 appears:3 ok:1 formulation:2 done:1 leen:1 vsv:13 just:3 smola:4 hand:1 horizontal:3 nonlinear:14 ild:1 propagation:1 defines:1 normalized:1 true:1 equality:1 hence:1 regularization:1 read:3 deal:2 width:1 complete:1 fj:1 aiyi:1 image:3 recently:1 superior:2 rotation:3 common:1 functional:2 jp:1 volume:3 extend:3 belong:1 significant:1 cambridge:1 ai:1 had:1 dot:2 chapelle:4 longer:1 add:3 showed:1 binary:1 yi:3 muller:2 seen:1 aut:2 signal:1 ii:1 full:1 multiple:2 lai:1 basic:1 iteration:1 kernel:30 achieved:1 justified:1 whereas:2 want:3 victorri:1 leaving:1 diagonalize:3 rest:1 probably:1 subject:1 dxj:3 call:1 extracting:1 split:3 xj:13 affect:1 inner:1 idea:3 multiclass:1 translates:1 expression:1 pca:4 bartlett:1 ird:1 svs:1 yik:1 useful:1 generally:2 extensively:1 concentrated:1 svms:8 notice:1 sign:2 per:1 neither:1 sum:1 enforced:2 extends:1 almost:1 reasonable:1 vn:1 decision:8 adapted:1 constraint:3 orthogonality:1 precisely:1 x2:2 bousquet:1 speed:1 span:2 fayyad:1 ltx:1 llj:2 according:1 belonging:1 across:1 son:1 presently:1 projecting:1 invariant:12 explained:1 equation:3 previously:1 turn:1 know:2 enforcement:2 merit:1 end:1 denker:1 uthurusamy:1 alternative:2 original:3 top:1 cf:2 especially:1 already:2 biowulf:2 aiajyiyjk:1 diagonal:1 unclear:1 gradient:1 subspace:1 distance:2 mapped:2 separating:1 berlin:1 argue:1 tuebingen:1 trivial:1 reason:7 enforcing:1 length:1 minimizing:1 sigma:3 perform:1 sch6lkopf:8 vertical:4 benchmark:1 finite:3 nist:1 descent:1 tok:1 november:1 extended:2 community:1 overcoming:1 introduced:1 namely:1 paris:1 learned:1 nip:2 able:1 suggested:1 bar:2 pattern:1 max:1 difficulty:1 hybrid:2 regularized:1 improve:1 technology:2 axis:1 prior:3 discovery:1 tangent:16 lecture:1 permutation:1 foundation:1 eigendecomposition:2 degree:1 article:1 editor:6 translation:5 lo:1 keeping:1 side:3 burges:4 institute:1 taking:1 boundary:5 dimension:4 xn:4 gram:3 world:1 valid:1 made:1 preprocessing:1 far:2 tiibingen:1 bernhard:2 xi:38 why:2 promising:1 schuurmans:1 expansion:2 necessarily:1 constructing:1 did:1 icann:1 main:1 whole:2 coordinatewise:1 allowed:2 repeated:1 en:1 wiley:1 dlj:5 kfj:1 xl:3 lie:1 svm:18 cortes:1 incorporating:5 vapnik:6 adding:4 margin:6 kx:1 lt:2 led:1 expressed:1 springer:2 corresponds:2 ma:2 conditional:1 rbf:2 change:2 hard:1 included:1 uniformly:1 hyperplane:4 principal:3 called:4 invariance:26 experimental:1 support:20 latter:2 incorporate:2 phenomenon:1
1,124
2,025
A Model of the Phonological Loop: Generalization and Binding Randall C. O'Reilly Department of Psychology University of Colorado Boulder 345 UCB Boulder, CO 80309 Rodolfo Soto Department of Psychology University of Colorado Boulder 345 UCB Boulder, CO 80309 oreilly@psych.colorado.edu Abstract We present a neural network model that shows how the prefrontal cortex, interacting with the basal ganglia, can maintain a sequence of phonological information in activation-based working memory (i.e., the phonological loop). The primary function of this phonological loop may be to transiently encode arbitrary bindings of information necessary for tasks - the combinatorial expressive power of language enables very flexible binding of essentially arbitrary pieces of information. Our model takes advantage of the closed-class nature of phonemes, which allows different neural representations of all possible phonemes at each sequential position to be encoded. To make this work, we suggest that the basal ganglia provide a region-specific update signal that allocates phonemes to the appropriate sequential coding slot. To demonstrate that flexible, arbitrary binding of novel sequences can be supported by this mechanism, we show that the model can generalize to novel sequences after moderate amounts of training. 1 Introduction Sequential binding is a version of the binding problem requiring that the identity of an item and its position within a sequence be bound. For example, to encode a phone number (e.g., 492-0054), one must remember not only the digits, but their order within the sequence. It has been suggested that the brain may have developed a specialized system for this form of binding in the domain of phonological sequences, in the form of the phonological loop (Baddeley, 1986; Baddeley, Gathercole, & Papagno, 1998; Burgess & Hitch, 1999). The phonological loop is generally conceived of as a system that can quickly encode a sequence of phonemes and then repeat this sequence back repeatedly. Standard estimates place the capacity of this loop at about 2.5 seconds of "inner speech," and it is widely regarded as depending on the prefrontal cortex (e.g. , Paulesu, Frith, & Frackowiak, 1993). We have developed a model of the phonological loop based on our existing framework for understanding how the prefrontal cortex and basal ganglia interact to support activation-based working memory (Frank, Loughry, & O'Reilly, 2001). This model performs binding by using different neural substrates for the different sequential positions of phonemes. This is a viable solution for a small, closed-class set of items like phonemes. However, through the combinatorial power of language, these phonological sequences can represent a huge number of distinct combinations of concepts. Therefore, this basic maintenance mechanism can be leveraged in many different circumstances to bind information needed for immediate use (e.g., in working memory tasks). A good example of this form of transient, phonologically-dependent binding comes from a task studied by Miyake and Soto (in preparation). In this task, participants saw sequentially-presented colored letters one at a time on a computer display, and had to respond to targets of a red X or a green Y, but not to any other color-letter combination (e.g., green X's and red Y's, which were also presented). After an initial series of trials with this set of targets, the targets were switched to be a green X and a red Y. Thus, the task clearly requires binding of color and letter information, and updating of these bindings after the switch condition. Miyake and Soto (in preparation) found that if they simply had participants repeat the word "the" over and over during the task (i.e., articulatory suppression), it interfered significantly with performance. In contrast, performing a similar repeated motor response that did not involve the phonological system (repeated foot tapping) did not interfere (but this task did interfere at the same level as articulatory suppression in a control visual search task, so one cannot argue that the interference was simply a matter of differential task difficulty). Miyake and Soto (in preparation) interpret this pattern of results as showing that the phonological loop supports the binding of stimulus features (e.g., participants repeatedly say to themselves "red X, green y' .. " , which is supported by debriefing reports), and that the use of this phonological system for unrelated information during articulatory suppression leads to the observed performance deficits. This form of phonological binding can be contrasted with other forms of binding that can be used in other situations and subserved by other brain areas besides the prefrontal cortex. O'Reilly, Busby, and Soto (in press) identify two other important binding mechanisms and their neural substrates in addition to the phonological loop mechanism: ? Cortical coarse-coded conjunctive binding: This is where each neural unit codes in a graded fashion for a large number of relatively low-order conjunctions, and many such units are used to represent any given input (e.g., Wickelgren, 1969; Mel & Fiser, 2000; O'Reilly & Busby, 2002). This form of binding takes place within the basic representations in the network that are shaped by gradual learning processes and provides a long-lasting (nontransient) form of binding. In short, these kinds of distributed representations avoid the binding problem in the first place by ensuring that relevant conjunctions are encoded, instead of representing different features using entirely separate, localist units (which is what gives rise to binding problems in the first place). However, this form of binding cannot rapidly encode novel bindings required for specific tasks - the phonological loop mechanism can thus complement the basic cortical mechanism by providing flexible, transient bindings on an ad-hoc basis. ? Hippocampal episodic conjunctive binding: Many theories of hippocampal function converge on the idea that it binds together individual elements of an experience into a unitary representation, which can for example be later recalled from partial cues (see O'Reilly & Rudy, 2001 for a review). These hippocampal conjunctive representations are higher-order and more spe- cific than the lower-order coarse-coded cortical conjunctive representations (i.e., a hippocampal conjunction encodes the combination of many feature elements, while a cortical conjunction encodes relatively few). Thus, the hippocampus can be seen as a specialized system for doing long-term binding of specific episodes, complementing the more generalized conjunctive binding performed by the cortex. Importantly, the hippocampus can also encode these conjunctions rapidly, and therefore it shares some of the same functionality as the phonological loop mechanism (i.e., rapidly encoding arbitrary conjunctions required for tasks). Thus, it is likely that the hippocampus and the prefrontal-mediated working memory system (including the phonological loop) are partially redundant with each other, and work together in many tasks (Cohen & O'Reilly, 1996). 2 Prefrontal Cortex and Basal Ganglia in Working Memory Our model of the phonological loop takes advantage of recent work showing how the prefrontal cortex and basal ganglia can interact to support activation-based working memory (Frank et al., 2001). The critical principles behind this work are as follows: ? Prefrontal cortex (PFC) is specialized relative to the posterior cortex for robust and rapidly updatable maintenance of information in an active state (i.e., via persistent firing of neurons). Thus, PFC can quickly update to maintain new information (in this case, the one exposure to a sequence of phonemes), while being able to also protect maintained information from interference from ongoing processing (see O'Reilly, Braver, & Cohen, 1999; Cohen, Braver, & O'Reilly, 1996; Miller & Cohen, 2001 for elaborations and reviews of relevant data). ? Robust maintenance and rapid updating are in fundamental conflict, and require a dynamic gating mechanism that can switch between these two modes of operation (O'Reilly et al., 1999; Cohen et al., 1996). ? The basal ganglia (BG) can provide this dynamic gating mechanism via modulatory, dis inhibitory connectivity with the PFC. Furthermore, this BG-based gating mechanism provides selectivity, such that separate regions of the PFC can be independently updated or allowed to perform robust maintenance. A possible anatomical substrate for these separably updatable PFC regions are the stripe structures identified by Levitt, Lewis, Yoshioka, and Lund (1993). ? Active maintenance in the PFC is implemented via a combination of recurrent excitatory connections and intracellular excitatory ionic conductances. This allows the PFC units to generally reflect the current inputs, except when these units have their intracellular maintenance currents activated, which causes them to reflect previously maintained information. See Frank et al. (2001) for more details on the importance of this mechanism. 3 Phonological Loop Model The above mechanisms motivated our modeling of the phonological loop as follows (see Figure 1) . First, separate PFC stripes are used to encode each step in the sequence. Thus, binding of phoneme identity and sequential order occurs in this model by using distinct neural substrates to represent the sequential information. This is entirely feasible because each stripe can represent all of the possible phonemes, given that they represent a closed class of items. Second, the storage of a Figure 1: Phonological loop model. Ten different input symbols are possible at each time step (one unit out of ten activated in the Input layer) . A sequence is encoded in one pass by presenting the Input together with the sequential location in the Time input layer for each step in the sequence. The simulated basal ganglia gating mechanism (implemented by fiat in script code) uses the time input to trigger intracellular maintenance currents in the corresponding stripe region of the context (PFC) layer (stripes are shown as the three separate groups of units within the Context layer; individual context units also had an excitatory self-connection for maintenance). Thus, the first stripe must learn to encode the first input, etc. Immediately after encoding, the network is then trained to produce the correct output in response to the time input, without any Input activation (the activation state shown is the network correctly recalling the third item in a sequence). The hidden layer must therefore learn to decode the context representations for this recall phase. Generalization testing involved presenting untrained sequences. new sequence involves the basal ganglia gating mechanism triggering updates of the different PFC stripes in the appropriate order. We assume this can be learned over experience, and we are currently working on developing powerful learning mechanisms for adapting the basal ganglia gating mechanism in this way. This kind of gating control would also likely require some kind of temporal/sequential input that indicates the location within the sequence - such information might come from the cerebellum (e.g., Ivry, 1996). In advance of having developed realistic and computationally powerful mechanisms for both the learning and the temporal/sequential control aspects of the model, we simply implemented these by fiat in the simulator. For the temporal signal indicating location within the sequence, we simply activated a different individual time unit for each point in the sequence (the Time input layer in Figure 1). This signal was then used by a simulated gating mechanism (implemented in script code in the simulator) to update the corresponding stripe in prefrontal cortex. Although the resulting model was therefore simplified, it nevertheless still had a challenging learning task to perform. Specifically, the stripe context layers had to learn to encode and maintain the current input value properly, and the Hidden layer had to be able to decode the context layer information as a function of the time input value. The model was implemented using the Leabra algorithm with standard parameters (O'Reilly, 1998; O 'Reilly & Munakata, 2000). Phonological Loop Generalization ..e 0.3 ali 0.2 I: .S! "lii .. .!::! iij Q) 0.1 I: Q) Cl 0.0 100 200 300 800 Number of Training Events Figure 2: Generalization results for the phonological loop model as a function of number training patterns. Generalization is over 90% correct with training on less than 20% of the possible input patterns. N = 5. 3 .1 Network Training The network was trained as follows. Sequences (of length 3 for our initial work) were presented by sequentially activating an input "phoneme" and a corresponding sequential location input (in the Time input layer) . We only used 10 different phonemes, each of which was encoded locally with a different unit in the Input layer. For example, the network could get Time = 0, Input = 2, then Time = 1, Input = 7, then Time = 2, Input = 3 to encode the sequence 2,7,3. During this encoding phase, the network was trained to activate the current Input on the Output layer, and the simulated gating function simply activated the intracellular maintenance currents for the units in the stripe in the Context (PFC) layer that corresponded to the Time input (i.e., stripe 0 for Time=O, etc). Then, the network was trained to recall this sequence, during which t ime no Input activation was present. The network received the sequence of Time inputs (0,1,2), and was trained to produce the corresponding Output for that location in the sequence (e.g., 2,7,3). The PFC context layers just maintained their activation states based on the intracellular ion currents activated during encoding (and recurrent activation) - once the network has been trained, the active PFC state represents the entire sequence. 3.2 Generalization Results A critical test of the model is to determine whether it can perform systematically with novel sequences - only if it demonstrates this capacity can it serve as a mechanism for rapidly binding arbitrary information (such as the task demands studied by Miyake & Soto, in preparation). With 10 input phonemes and sequences of length t hree, there were 1,000 different sequences possible (we allowed phonemes to repeat). We trained on 100, 200, 300, and 800 of these sequences, and tested generalization on the remaining sequences. The generalization results are shown in Figure 2, which clearly shows that the network learned these sequences in a systematic manner and could transfer its training knowledge to novel sequences. Interestingly, there appears to be a critical transition between 100 and 200 training sequences - 100 sequences corresponds to each item within each slot being presented roughly 10 times, which appears to provide sufficient statistical information regarding the independence of individual slots. Figure 3: Hidden unit representations (values are weights into a hidden unit from all other layers). Unit in a) encodes the conjunction of a subset of input/output items at time 2. (b) encodes a different subset of items at time 2. (c) encodes items over times 2 and 3. (d) has no selectivity in the input, but does project to the output and likely participates in recall of items at time step 3. 3.3 Analysis of Representations To understand how the hidden units encode and retrieve information in the maintained context layer in a systematic fashion that supports the good generalization observed, we examined the patterns of learned weights. Some representative examples are shown in Figure 3. Here, we see evidence of coarse-coded representations that encode a subset of items in either one time point in the sequence or a couple of time points. Also we found units that were more clearly associated with retrieval and not encoding. These types of representations are consistent with our other work showing how these kinds of representations can support good generalization (O'Reilly & Busby, 2002). 4 Discussion We have presented a model of sequential encoding of phonemes, based on independently-motivated computational and biological considerations, focused on the neural substrates of the prefrontal cortex and basal ganglia (Frank et al., 2001). Viewed in more abstract, functional terms, however , our model is just another in a long line of computational models of how people might encode sequential order information. There are two classic models: (a) associative chaining, where the acti- vation of a given item triggers the activation of the next item via associative links, and (b) item-position association models where items are associated with their sequential positions and recalled from position cues (e.g., Lee & Estes, 1977). The basic associative chaining model has been decisively ruled out based on error patterns (Henson, Norris, Page, & Baddeley, 1996), but modified versions of it may avoid these problems (e.g., Lewandowsky & Murdock, 1989). Probably the most accomplished current model, Burgess and Hitch (1999), is a version of the itemposition association model with a competitive queuing mechanism where the most active item is output first and is then suppressed to allow other items to be output. Compared to these existing models, our model is unique in not requiring fast associational links to encode items within the sequence. For example, the Burgess and Hitch (1999) model uses rapid weight changes to associate items with a context representation that functions much like the time input in our model. In contrast, items are maintained strictly via persistent activation in our model , and the basalganglia based gating mechanism provides a means of encoding items into separate neural slots that implicitly represent sequential order. Thus, the time inputs act independently on the basal ganglia, which then operates generically on whatever phoneme information is presently activated in the auditory input, obviating the need for specific item-context links. The clear benefit of not requiring associationallinks is that it makes the model much more flexible and capable of generalization to novel sequences as we have demonstrated here (see O'Reilly & Munakata, 2000 for extended discussion of this general issue). Thus, we believe our model is uniquely well suited for explaining the role of the phonological loop in rapid binding of novel task information. Nevertheless, the present implementation of the model has numerous shortcomings and simplifications, and does not begin to approach the work of Burgess and Hitch (1999) in accounting for relevant psychological data. Thus, future work will be focused on remedying these limitations. One important issue that we plan to address is the interplay between the present model based on the prefrontal cortex and the binding that the hippocampus can provide - we suspect that the hippocampus will contribute item-position associations and their associated error patterns and other phenomena as discussed in Burgess and Hitch (1999). Acknowledgments This work was supported by ONR grant N00014-00-1-0246 and NSF grant IBN9873492. Rodolfo Soto died tragically at a relatively young age during the preparation of this manuscript - this work is dedicated to his memory. 5 References Baddeley, A. , Gathercole, S. , & Papagno, C. (1998). The phonological loop as a language learning device. Psychological Review, 105, 158. Baddeley, A. D. (1986). Working memory. New York: Oxford University Press. Burgess, N. , & Hitch, G. J . (1999). Memory for serial order: A network model of the phonological loop and its timing. Psychological Review, 106, 551- 581. Cohen, J. D., Braver, T. S., & O'Reilly, R. C. (1996). A computational approach to prefrontal cortex, cognitive control, and schizophrenia: Recent developments and current challenges. Philosophical Transactions of the Royal Society (London) B, 351, 1515- 1527. Cohen, J. D. , & O'Reilly, R. C. (1996). A preliminary theory of the interactions between prefrontal cortex and hippocampus that contribute to planning and prospective memory. In M. Brandimonte, G. O. Einstein, & M. A. McDaniel (Eds.) , Prospective memory: Theory and applications (pp. 267- 296). Mahwah, New Jersey: Erlbaum. Frank, M. J. , Loughry, B., & O 'Reilly, R . C. (2001). Interactions between the frontal cortex and basal ganglia in working memory: A computational model. Cognitive, Affective, and Behavioral Neurosci ence, 1 , 137- 160. Henson, R. N. A., Norris, D. G ., Page, M. P . A., & Baddeley, A. D . (1996) . Unclaimed memory: Error patterns rule out chaining models of immediate serial recall. Quarterly Journal of Experimental Psychology: Human Experim ental Psychology, 49(A) , 80- 115. Ivry, R. (1996). The representation of temporal information in perception and motor control. Current Opinion in N eurobiology, 6,851-857. Lee, C. L. , & Estes, W. K. (1977). Order and position in primary memory for letter strings. Journal of Verbal Learning and Verbal B ehavior, 16, 395- 418. Levitt , J . B. , Lewis, D. A. , Yoshioka, T. , & Lund, J. S. (1993). Topography of pyramidal neuron intrinsic connections in macaque monkey prefrontal cortex (areas 9 & 46). Journal of Comparativ e N eurology, 338, 360- 376. Lewandowsky, S. , & Murdock, B. B. (1989). Memory for serial order. Psychological R eview, 96, 25- 57. Mel, B. A., & Fiser, J. (2000). Minimizing binding errors using learned conjunctive features. Neural Computation, 12, 731- 762. Miller, E. K. , & Cohen, J. D. (2001). An integrative theory of prefrontal cortex function. Annual Review of Neuroscience , 24, 167- 202. Miyake, A., & Soto , R. (in preparation). The role of the phonological loop in executive control. O'Reilly, R. C. (1998). Six principles for biologically-based computational models of cortical cognition. Trends in Cognitive Sci ences, 2(11), 455- 462. O 'Reilly, R . C ., Braver, T . S., & Cohen, J . D . (1999) . A biologically based computational model of working memory. In A. Miyake, & P. Shah (Eds.) , Mod els of working m emory: M echanisms of active maintenance and executiv e control. (pp. 375- 411) . New York: Cambridge University Press. O 'Reilly, R. C. , & Busby, R. S. (2002). Generalizable relational binding from coarsecoded distributed representations. Advances in N eural Information Processing Systems (NIPS), 2001. O 'Reilly, R . C. , Busby, R. S., & Soto, R. (in press). Three forms of binding and their neural substrates: Alternatives to temporal synchrony. In A. Cleeremans (Ed.) , Th e unity of consciousness: Binding, integration, and dissociation. Oxford: Oxford University Press. O 'Reilly, R. C ., & Munakata, Y . (2000) . Computational explorations in cognitive n euroscience: Understanding th e mind by simulating th e brain. Cambridge, MA: MIT Press. O 'Reilly, R. C. , & Rudy, J. W . (2001). Conjunctive representations in learning and memory: Principles of cortical and hippocampal function . Psychological Review, 108, 311345. Paulesu, E. , Frith, C. D. , & Frackowiak, R. S. J. (1993). The neural correlates of the verbal component of working memory. Nature, 362,342- 345. Wickelgren, W. A. (1969). Context-sensitive coding, associative memory, and serial order in (speech) behavior. Psychological R eview, 76 , 1- 15 .
2025 |@word trial:1 version:3 hippocampus:6 ences:1 integrative:1 gradual:1 accounting:1 initial:2 series:1 interestingly:1 existing:2 current:10 emory:1 activation:10 conjunctive:7 must:3 realistic:1 enables:1 motor:2 update:4 cue:2 device:1 item:22 complementing:1 updatable:2 short:1 colored:1 provides:3 coarse:3 murdock:2 location:5 contribute:2 differential:1 viable:1 persistent:2 acti:1 affective:1 behavioral:1 manner:1 rapid:3 behavior:1 themselves:1 planning:1 roughly:1 simulator:2 brain:3 project:1 begin:1 unrelated:1 what:1 kind:4 psych:1 string:1 monkey:1 developed:3 generalizable:1 temporal:5 remember:1 act:1 demonstrates:1 control:7 unit:16 whatever:1 grant:2 bind:2 died:1 timing:1 encoding:7 oxford:3 tapping:1 firing:1 might:2 studied:2 examined:1 challenging:1 co:2 unique:1 acknowledgment:1 testing:1 loughry:2 digit:1 episodic:1 area:2 significantly:1 adapting:1 reilly:22 word:1 suggest:1 get:1 cannot:2 storage:1 context:12 paulesu:2 demonstrated:1 exposure:1 independently:3 focused:2 miyake:6 immediately:1 rule:1 regarded:1 importantly:1 his:1 retrieve:1 classic:1 updated:1 target:3 trigger:2 colorado:3 decode:2 substrate:6 us:2 associate:1 element:2 trend:1 updating:2 stripe:11 observed:2 role:2 cleeremans:1 region:4 episode:1 dynamic:2 trained:7 ali:1 serve:1 basis:1 frackowiak:2 jersey:1 distinct:2 fast:1 shortcoming:1 activate:1 london:1 corresponded:1 encoded:4 widely:1 say:1 associative:4 hoc:1 sequence:37 advantage:2 interplay:1 interaction:2 relevant:3 loop:22 rapidly:5 gathercole:2 produce:2 depending:1 recurrent:2 received:1 implemented:5 involves:1 come:2 foot:1 correct:2 functionality:1 exploration:1 human:1 transient:2 opinion:1 require:2 activating:1 generalization:11 preliminary:1 biological:1 strictly:1 cognition:1 combinatorial:2 currently:1 saw:1 sensitive:1 mit:1 clearly:3 modified:1 avoid:2 conjunction:7 encode:13 properly:1 indicates:1 contrast:2 suppression:3 yoshioka:2 dependent:1 el:1 entire:1 hidden:5 issue:2 flexible:4 development:1 plan:1 integration:1 eurobiology:1 once:1 phonological:27 shaped:1 having:1 represents:1 future:1 report:1 stimulus:1 transiently:1 few:1 ime:1 lewandowsky:2 individual:4 phase:2 maintain:3 recalling:1 conductance:1 decisively:1 huge:1 hitch:6 generically:1 behind:1 activated:6 articulatory:3 capable:1 partial:1 necessary:1 experience:2 allocates:1 ruled:1 psychological:6 modeling:1 ence:1 localist:1 subset:3 erlbaum:1 vation:1 rudy:2 fundamental:1 systematic:2 participates:1 lee:2 together:3 quickly:2 connectivity:1 reflect:2 leveraged:1 prefrontal:15 wickelgren:2 cognitive:4 lii:1 coding:2 matter:1 ad:1 bg:2 piece:1 queuing:1 later:1 performed:1 closed:3 script:2 doing:1 red:4 competitive:1 participant:3 synchrony:1 phoneme:15 miller:2 dissociation:1 identify:1 generalize:1 ionic:1 coarsecoded:1 ed:3 pp:2 involved:1 spe:1 phonologically:1 associated:3 couple:1 auditory:1 recall:4 color:2 knowledge:1 oreilly:1 fiat:2 eview:2 back:1 appears:2 manuscript:1 higher:1 response:2 furthermore:1 just:2 fiser:2 working:12 expressive:1 interfere:2 mode:1 soto:9 believe:1 requiring:3 concept:1 ivry:2 consciousness:1 cerebellum:1 during:6 self:1 uniquely:1 maintained:5 mel:2 chaining:3 generalized:1 busby:5 hippocampal:5 presenting:2 demonstrate:1 performs:1 dedicated:1 consideration:1 novel:7 specialized:3 functional:1 cohen:9 association:3 discussed:1 interpret:1 cambridge:2 language:3 had:6 henson:2 cortex:17 etc:2 posterior:1 recent:2 moderate:1 phone:1 selectivity:2 n00014:1 onr:1 accomplished:1 seen:1 converge:1 determine:1 redundant:1 signal:3 long:3 retrieval:1 elaboration:1 serial:4 schizophrenia:1 coded:3 ensuring:1 basic:4 maintenance:10 essentially:1 circumstance:1 represent:6 ion:1 addition:1 pyramidal:1 probably:1 suspect:1 mod:1 unitary:1 switch:2 independence:1 remedying:1 psychology:4 burgess:6 identified:1 triggering:1 inner:1 idea:1 regarding:1 whether:1 motivated:2 cific:1 six:1 speech:2 york:2 cause:1 repeatedly:2 generally:2 modulatory:1 clear:1 involve:1 amount:1 ten:2 locally:1 mcdaniel:1 nsf:1 inhibitory:1 neuroscience:1 conceived:1 correctly:1 anatomical:1 basal:12 group:1 nevertheless:2 letter:4 powerful:2 respond:1 place:4 entirely:2 bound:1 layer:16 simplification:1 display:1 annual:1 encodes:5 aspect:1 performing:1 relatively:3 rodolfo:2 department:2 developing:1 leabra:1 combination:4 suppressed:1 unity:1 biologically:2 randall:1 lasting:1 presently:1 boulder:4 interference:2 computationally:1 previously:1 mechanism:21 needed:1 mind:1 operation:1 experim:1 quarterly:1 einstein:1 appropriate:2 simulating:1 braver:4 alternative:1 shah:1 hree:1 remaining:1 estes:2 graded:1 society:1 occurs:1 primary:2 deficit:1 separate:5 separably:1 capacity:2 simulated:3 link:3 sci:1 prospective:2 argue:1 besides:1 code:3 length:2 providing:1 tragically:1 minimizing:1 frank:5 rise:1 implementation:1 perform:3 neuron:2 immediate:2 situation:1 extended:1 relational:1 interacting:1 arbitrary:5 complement:1 required:2 connection:3 philosophical:1 conflict:1 recalled:2 learned:4 protect:1 macaque:1 nip:1 address:1 able:2 suggested:1 pattern:7 perception:1 lund:2 challenge:1 green:4 memory:19 including:1 royal:1 power:2 critical:3 event:1 difficulty:1 representing:1 numerous:1 mediated:1 review:6 understanding:2 relative:1 topography:1 limitation:1 age:1 executive:1 switched:1 sufficient:1 consistent:1 principle:3 systematically:1 share:1 excitatory:3 supported:3 repeat:3 dis:1 verbal:3 allow:1 understand:1 explaining:1 distributed:2 benefit:1 cortical:6 transition:1 ental:1 simplified:1 transaction:1 correlate:1 implicitly:1 sequentially:2 active:5 search:1 euroscience:1 subserved:1 nature:2 learn:3 robust:3 transfer:1 frith:2 interact:2 untrained:1 pfc:13 cl:1 domain:1 did:3 intracellular:5 neurosci:1 mahwah:1 repeated:2 allowed:2 obviating:1 eural:1 levitt:2 representative:1 fashion:2 iij:1 position:8 third:1 young:1 specific:4 showing:3 gating:10 symbol:1 evidence:1 intrinsic:1 sequential:14 importance:1 associational:1 interfered:1 demand:1 suited:1 simply:5 likely:3 ganglion:12 visual:1 partially:1 binding:34 norris:2 corresponds:1 lewis:2 ma:1 slot:4 identity:2 viewed:1 feasible:1 change:1 specifically:1 except:1 contrasted:1 operates:1 pas:1 experimental:1 munakata:3 ucb:2 indicating:1 support:5 people:1 frontal:1 preparation:6 ongoing:1 baddeley:6 tested:1 phenomenon:1
1,125
2,026
Modeling Temporal Structure in Classical Conditioning Aaron C. Courville 1 ,3 and David S. Touretzk y 2,3 1 Robotics Institute, 2Computer Science Department 3Center for the Neural Basis of Cognition Carnegie Mellon University, Pittsburgh, PA 15213-3891 { aarone, dst} @es.emu.edu Abstract The Temporal Coding Hypothesis of Miller and colleagues [7] suggests that animals integrate related temporal patterns of stimuli into single memory representations. We formalize this concept using quasi-Bayes estimation to update the parameters of a constrained hidden Markov model. This approach allows us to account for some surprising temporal effects in the second order conditioning experiments of Miller et al. [1 , 2, 3], which other models are unable to explain. 1 Introduction Animal learning involves more than just predicting reinforcement. The well-known phenomena of latent learning and sensory preconditioning indicate that animals learn about stimuli in their environment before any reinforcement is supplied. More recently, a series of experiments by R. R. Miller and colleagues has demonstrated that in classical conditioning paradigms, animals appear to learn the temporal structure of the stimuli [8]. We will review three of these experiments. We then present a model of conditioning based on a constrained hidden Markov model , using quasiBayes estimation to adjust the model parameters online. Simulation results confirm that the model reproduces the experimental observations, suggesting that this approach is a viable alternative to earlier models of classical conditioning which cannot account for the Miller et al. experiments. Table 1 summarizes the experimental paradigms and the results. Expt. 1: Simultaneous Conditioning. Responding to a conditioned stimulus (CS) is impaired when it is presented simultaneously with the unconditioned stimulus (US) rather than preceding the US. The failure of the simultaneous conditioning procedure to demonstrate a conditioned response (CR) is a well established result in the classical conditioning literature [9]. Barnet et al. [1] reported an interesting Expt. 1 Phase 1 (4)T+ US Phase 2 (4)C -+ T Expt.2A Expt. 2B (12)T -+ C (12)T -+ C (8)T -+ US (8)T ---+ US C=> C =>CR Expt. 3 (96)L -+ US -+ X (8) B -+ X X=> - Test => Result T=> - Test => Result C =>CR B =>CR Table 1: Experimental Paradigms. Phases 1 and 2 represent two stages of training trials, each presented (n) times. The plus sign (+ ) indicates simultaneous presentation of stimuli; the short arrow (-+) indicates one stimulus immediately following another; and the long arrow (-----+) indicates a 5 sec gap between stimulus offset and the following stimulus onset. For Expt. 1, the tone T, click train C, and footshock US were all of 5 sec duration. For Expt. 2, the tone and click train durations were 5 sec and the footshock US lasted 0.5 sec. For Expt. 3, the light L , buzzer E , and auditory stimulus X (either a tone or white noise) were all of 30 sec duration, while the footshock US lasted 1 sec. CR indicates a conditioned response to the test stimulus. second-order extension of the classic paradigm. While a tone CS presented simultaneously with a footshock results in a minimal CR to the tone, a click train preceding the tone (in phase 2) does acquire associative strength, as indicated by a CR. Expt. 2: Sensory Preconditioning. Cole et al. [2] exposed rats to a tone T immediately followed by a click train C. In a second phase, the tone was paired with a footshock US that either immediately followed tone offset (variant A), or occurred 5 sec after tone offset (variant B). They found that when C and US both immediately follow T , little conditioned response is elicited by the presentation of C. However, when the US occurs 5 sec after tone offset, so that it occurs later than C (measured relative to T), then C does come to elicit a CR. Expt. 3: Backward Conditioning. In another experiment by Cole et al. [3], rats were presented with a flashing light L followed by a footshock US, followed by an auditory stimulus X (either a tone or white noise). In phase 2, a buzzer B was followed by X. Testing revealed that while X did not elicit a CR (in fact, it became a conditioned inhibitor), X did impart an excitatory association to B. 2 Existing Models of Classical Conditioning The Rescorla-Wagner model [11] is still the best-known model of classical conditioning, but as a trial-level model, it cannot account for within-trial effects such as second order conditioning or sensitivity to stimulus timing. Sutton and Barto developed V-dot theory [14] as a real-time extension of Rescorla-Wagner. Further refinements led to the Temporal Difference (TD) learning algorithm [14]. These extensions can produce second order conditioning. And using a memory buffer representation (what Sutton and Barto call a complete serial compound), TD can represent the temporal structure of a trial. However, TD cannot account for the empirical data in Experiments 1- 3 because it does not make inferences about temporal relationships among stimuli; it focuses solely on predicting the US. In Experiment 1, some versions of TD can account for the reduced associative strength of a CS when its onset occurs simultaneously with the US, but no version of TD can explain why the second-order stimulus C should acquire greater associative strength than T. In Experiment 2, no learning occurs in Phase 1 with TD because no prediction error is generated by pairing T with C. As a result, no CR is elicited by C after T has been paired with the US in Phase 2. In Experiment 3, TD fails to predict the results because X is not predictive of the US; thus X acquires no associative strength to pass on to B in the second phase. Even models that predict future stimuli have trouble accounting for Miller et al. 's results. Dayan's "successor representation" [4], the world model of Sutton and Pinette [15], and the basal ganglia model of Suri and Schultz [13] all attempt to predict future stimulus vectors. Suri and Schultz's model can even produce one form of sensory preconditioning. However, none of these models can account for the responses in any of the three experiments in Table 1, because they do not make the necessary inferences about relations among stimuli. Temporal Coding Hypothesis The temporal coding hypothesis (TCH) [7] posits that temporal contiguity is sufficient to produce an association between stimuli. A CS does not need to predict reward in order to acquire an association with the US. Furthermore, the association is not a simple scalar quantity. Instead, information about the temporal relationships among stimuli is encoded implicitly and automatically in the memory representation of the trial. Most importantly, TCH claims that memory representations of trials with similar stimuli become integrated in such a way as to preserve the relative temporal information [3]. If we apply the concept of memory integration to Experiment 1, we get the memory representation, C ---+ T + US. If we interpret a CR as a prediction of imminent reinforcement, then we arrive at the correct prediction of a strong response to C and a weak response to T. Integrating the hypothesized memory representations of the two phases of Experiment 2 results in: A) T ---+ C+US and B) T ---+ C ---+ US. The stimulus C is only predictive ofthe US in variant B, consistent with the experimental findings. For Experiment 3, an integrated memory representation of the two phases produces L+ B ---+ US ---+ X. Stimulus B is predictive of the US while X is not. Thus, the temporal coding hypothesis is able to account for the results of each of the three experiments by associating stimuli with a timeline. 3 A Computational Model of Temporal Coding A straightforward formalization of a timeline is a Markov chain of states. For this initial version of our model, state transitions within the chain are fixed and deterministic. Each state represents one instant of time, and at each timestep a transition is made to the next state in the chain. This restricted representation is key to capturing the phenomena underlying the empirical results. Multiple timelines (or Markov chains) emanate from a single holding state. The transitions out of this holding state are the only probabilistic and adaptive transitions in the simplified model. These transition probabilities represent the frequency with which the timelines are experienced. Figure 1 illustrates the model structure used in all simulations. Our goal is to show that our model successfully integrates the timelines of the two training phases of each experiment. In the context of a collection of Markov chains, integrating timelines amounts to both phases of training becoming associated with a single Markov chain. Figure 1 shows the integration of the two phases of Expt. 2B. Figure 1: A depiction of the state and observation structure of the model. Shown are two timelines, one headed by state j and the other headed by state k. State i, the holding state, transitions to states j and k with probabilities aij and aik respectively. Below the timeline representations are a sequence of observations represented here as the symbols T, C and US. The T and C stimuli appear for two time steps each to simulate their presentation for an extended duration in the experiment. During the second phase of the experiments, the second Markov chain (shown in Figure 1 starting with state k) offers an alternative to the chain associated with the first phase of learning. If we successfully integrate the timelines, this second chain is not used. As suggested in Figure 1, associated with each state is a stimulus observation. "Stimulus space" is an n-dimensional continuous space, where n is the number of distinct stimuli that can be observed (tone, light, shock, etc.) Each state has an expectation concerning the stimuli that should be observed when that state is occupied. This expectation is modeled by a probability density function, over this space, defined by a mixture of two multivariate Gaussians. The probability density at stimulus observation xt in state i at time tis , where Wi is a mixture coefficient for the two Gaussians associated with state i. The Gaussian means /tiD and /til and variances ufo and ufl are vectors of the same dimension as the stimulus vector xt. Given knowledge of the state, the stimulus components are assumed to be mutually independent (covariance terms are zero). We chose a continuous model of observations over a discrete observation model to capture stimulus generalization effects. These are not pursued in this paper. For each state, the first Gaussian pdf is non-adaptive, meaning /tiO is fixed about a point in stimulus space representing the absence of stimuli. ufo is fixed as well. For the second Gaussian, /til and Ufl are adaptive. This mixture of one fixed and one adaptive Gaussian is an approximation to the animal's belief distribution about stimuli, reflecting the observed tolerance animals have to absent expected stimuli. Put another way, animals seem to be less surprised by the absence of an expected stimulus than by the presence of an unexpected stimulus. We assume that knowledge of the current state st is inaccessible to the learner. This information must be inferred from the observed stimuli. In the case of a Markov chain, learning with hidden state is exactly the problem of parameter estimation in hidden Markov models. That is, we must update the estimates of w, /tl and for ur each state, and aij for each state transition (out of the holding state), in order to maximize the likelihood of the sequence of observations The standard algorithm for hidden Markov model parameter estimation is the Baum-Welch method [10]. Baum-Welch is an off-line learning algorithm that requires all observations used in training to be held in memory. In a model of classical conditioning, this is an unrealistic assumption about animals' memory capabilities. We therefore require an online learning scheme for the hidden Markov model, with only limited memory requirements. Recursive Bayesian inference is one possible online learning scheme. It offers the appealing property of combining prior beliefs about the world with current observations through the recursive application of Bayes' theorem, p(Alxt) IX p(xt lx t - 1 , A)p(AIXt - 1 ). The prior distribution, p(AIX t - 1 ) reflects the belief over the parameter A before the observation at time t , xt. X t - 1 is the observation history up to time t - l , i.e. X t - 1 = {x t - 1 ,xt - 2 , ... }. The likelihood, p(xtIXt-l,A) is the probability density over xt as a function of the parameter A. Unfortunately, the implementation of exact recursive Bayesian inference for a continuous density hidden Markov model (CDHMM) is computationally intractable. This is a consequence of there being missing data in the form of hidden state. With hidden state, the posterior distribution over the model parameters, after the observation, is given by N p(Alxt) IX LP(xtlst = i, X t - 1 , A)p(st = iIX t - 1 , A)p(AIXt - 1 ), (2) i=1 where we have summed over the N hidden states. Computing the recursion for multiple time steps results in an exponentially growing number of terms contributing to the exact posterior. We instead use a recursive quasi-Bayes approximate inference scheme developed by Huo and Lee [5], who employ a quasi-Bayes approach [12]. The quasi-Bayes approach exploits the existence of a repeating distribution (natural conjugate) over the parameters for the complete-data CDHMM. (i.e. where missing data such as the state sequence is taken to be known). Briefly, we estimate the value of the missing data. We then use these estimates, together with the observations, to update the hyperparameters governing the prior distribution over the parameters (using Bayes' theorem). This results in an approximation to the exact posterior distribution over CDHMM parameters within the conjugate family of the complete-data CDHMM. See [5] for a more detailed description of the algorithm. Estimating the missing data (hidden state) involves estimating transition probabilities between states, ~0 = Pr(sT = i, ST+1 = jlXt , A), and joint state and mixture component label probabilities ([k = Pr(sT = i, IT = klX t , A). Here zr = k is the mixture component label indicating which Gaussian, k E {a, I}, is the source of the stimulus observation at time T. A is the current estimate of all model parameters. We use an online version of the forward-backward algorithm [6] to estimate ~0 and ([1. The forward pass computes the joint probability over state occupancy (taken to be both the state value and the mixture component label) at time T and the sequence of observations up to time T. The backward pass computes the probability of the observations in a memory buffer from time T to the present time t given the state occupancy at time T. The forward and backward passes over state/observation sequences are combined to give an estimate of the state occupancy at time T given the observations up to the present time t. In the simulations reported here the memory buffer was 7 time steps long (t - T = 6). We use the estimates from the forward-backward algorithm together with the observations to update the hyperparameters. For the CDHMM, this prior is taken to be a product of Dirichlet probability density functions (pdfs) for the transition probabilities (aij) , beta pdfs for the observation model mixture coefficients (Wi) and normal-gamma pdfs for the Gaussian parameters (Mil and afl)' The basic hyperparameters are exponentially weighted counts of events, with recency weighting determined by a forgetting parameter p. For example, "'ij is the number of expected transitions observed from state i to state j, and is used to update the estimate of parameter aij. The hyperparameter Vik estimates the number of stimulus observations in state i credited to Gaussian k , and is used to update the mixture parameter Wi. The remaining hyperparameters 'Ij;, ?, and () serve to define the pdfs over Mil and afl' The variable d in the equations below indexes over stimulus dimensions. Si1d is an estimate of the sample variance, and is a constant in the present model. T _ ((T-1) "'ij - . I,T P "'ij . 1,(T-1) 'l' i1d = ()T _ p() ( T- 1) i1d - - P 'I' i1d i1d 1) (:T + 1 + '>ij T _ rT + '>i1 + 7" (i1 Sild 2 ((T-1) v ik - ,/,T P v ik _ p(,/,(T-1) _ 'l'i 1d - + 0,,( 7"-1 ) ,7" Po/ i 1d 'il 2(p1/Ji;d 1) H[1) (xT _ d - 'l'i1d 1) rT + 1 + '>ik 1) 2 + 1H[1 2 () II. T- 1 )2 f"'i 1d In the last step of our inference procedure, we update our estimate of the model parameters as the mode of their approximate posterior distribution. While this is an approximation to proper Bayesian inference on the parameter values, the mode of the approximate posterior is guaranteed to converge to a mode of the exact posterior. In the equations below, N is the number of states in the model. T_ Wi - 4 v[1- 1 vio + viI -2 Results and Discussion The model contained two timelines (Markov chains). Let i denote the holding state and j, k the initial states of the two chains. The transition probabilities were initialized as aij = aik = 0.025 and aii = 0.95. Adaptive Gaussian means Mild were initialized to small random values around a baseline of 10- 4 for all states. The exponential forgetting factor was P = 0.9975, and both the sample variances Si1d and the fixed variances aIOd were set to 0.2. We trained the model on each of the experimental protocols of Table 1, using the same numbers of trials reported in the original papers. The model was run continuously through both phases of the experiments with a random intertrial interval. '+-:::: noCR CR '+-:::: 4 5 '+ -:g4 noCR ~3 C OJ E OJ E OJ E ~3 .E ~2 OJ ~ () 02 .E c C Oi ~1 &!1 "Qi 0 a: g;1 0 trr trr T C Experiment 1 0 noCR trr (A)C (B)C Experiment 2 0 X B Experiment 3 Figure 2: Results from 20 runs of the model simulation with each experimental paradigm. On the ordinate is the total reinforcement (US) , on a log scale, above the baseline (an arbitrary perception threshold) expected to occur on the next time step. The error bars represent two standard deviations away from the mean. Figure 2 shows t he simulation results from each of the three experiments. If we assume that the CR varies monotonically with the US prediction, then in each case, t he model's predicted CR agreed with the observations of Miller et al. The CR predictions are the result of the model integrating t he two phases of learning into one t imeline. At the t ime of the presentation of the Phase 2 stimuli, the states forming the timeline describing the Phase 1 pattern of stimuli were judged more likely to have produced the Phase 2 stimuli than states in the other t imeline, which served as a null hypothesis. In another experiment, not shown here , we trained the model on disjoint stimuli in the two phases. In that situation it correctly chose a separate t imeline for each phase, rather than merging the two . We have shown that under the assumption t hat observation probabilities are modeled by a mixture of Gaussians, and a very restrictive state transition structure, a hidden Markov model can integrate the memory representations of similar temporal stimulus patterns. "Similarity" is formalized in this framework as likelihood under the t imeline model. We propose t his model as a mechanism for the integration of memory representations postulated in the Temporal Coding Hypothesis. The model can be extended in many ways. The current version assumes t hat event chains are long enough to represent an entire trial, but short enough that the model will return to the holding state before the start of the next trial. An obvious refinement would be a mechanism to dynamically adjust chain lengths based on experience. We are also exploring a generalization of the model to the semi-Markov domain, where state occupancy duration is modeled explicitly as a pdf. State transitions would then be tied to changes in observations, rather than following a rigid progression as is currently the case. Finally, we are experiment ing with mechanisms that allow new chains to be split off from old ones when the model determines that current stimuli differ consistently from t he closest matching t imeline. Fitting stimuli into existing t imelines serves to maximize the likelihood of current observations in light of past experience. But why should animals learn the temporal structure of stimuli as t imelines? A collection of timelines may be a reasonable model of the natural world. If t his is true, t hen learning with such a strong inductive bias may help t he animal to bring experience of related phenomena to bear in novel sit uations- a desirable characteristic for an adaptive system in a changing world. Acknowledgments Thanks to Nathaniel Daw and Ralph Miller for helpful discussions. This research was funded by National Science Foundation grants IRI-9720350 and IIS-997S403. Aaron Courville was funded in part by a Canadian NSERC PGS B fellowship. References [1] R. C. Barnet, H. M. Arnold, and R. R. Miller. Simultaneous conditioning demonstrated in second-order conditioning: Evidence for similar associative structure in forward and simultaneous conditioning. Learning and Motivation, 22:253- 268, 1991. [2] R. P. Cole, R. C. Barnet, and R. R . Miller. Temporal encoding in trace conditioning. Animal Learning and Behavior, 23(2) :144- 153, 1995 . [3] R. P. Cole and R. R. Miller. Conditioned excitation and conditioned inhibition acquired through backward conditioning. Learning and Motivation , 30:129- 156, 1999. [4] P. Dayan. Improving generalization for temporal difference learning: the successor representation. Neural Computation, 5:613- 624, 1993. [5] Q. Huo and C.-H. Lee. On-line adaptive learning of the continuous density hidden Markov model based on approximate recursive Bayes estimate. IEEE Transactions on Speech and Audio Processing, 5(2):161- 172, 1997. [6] V . Krishnamurthy and J . B. Moore. On-line estimation of hidden Markov model parameters based on the Kullback-Leibler information measure. IEEE Transactions on Signal Processing, 41(8):2557- 2573, 1993. [7] L. D. Matzel , F. P. Held, and R. R. Miller. Information and the expression of simultaneous and backward associations: Implications for contiguity theory. Learning and Motivation, 19:317- 344, 1988. [8] R. R. Miller and R . C. Barnet. The role of time in elementary associations. Current Directions in Psychological Sci ence, 2(4):106- 111 , 1993. [9] 1. P. Pavlov. Conditioned Reflexes. Oxford University Press, 1927. [10] L. R. Rabiner. A tutorial on hidden Markov models and selected applications speech recognition. Proceedings of th e IEEE, 77(2) :257- 285, 1989. III [11] R. A. Rescorla and A. R. Wagner. A theory of Pavlovian conditioning: Variations in the effectiveness of reinforcement and nonreinforcement . In A. H. Black and W. F. Prokasy, editors, Classical Conditioning II. Appleton-Century-Crofts, 1972. [12] A. F . M. Smith and U. E . Makov . A quasi-Bayes sequential procedure for mixtures. Journal of th e Royal Statistical Soci ety, 40(1):106- 112, 1978. [13] R. E. Suri and W. Schultz. Temporal difference model reproduces anticipatory neural activity. N eural Computation, 13(4):841- 862, 200l. [14] R. S. Sutton and A. G. Barto. Time-derivative models of Pavlovian reinforcement. In M. Gabriel and J. Moore, editors, Learning and Computational N euroscience: Foundations of Adaptive N etworks, chapter 12 , pages 497- 537. MIT Press, 1990. [15] R. S. Sutton and B. Pinette. The learning of world models by connectionist networks. In L. Erlbaum, editor, Proceedings of the seventh annual conference of the cognitive science society, pages 54- 64, Irvine, California, August 1985.
2026 |@word mild:1 trial:9 version:5 briefly:1 simulation:5 t_:1 accounting:1 covariance:1 initial:2 series:1 past:1 existing:2 current:7 surprising:1 must:2 afl:2 update:7 pursued:1 selected:1 tone:13 huo:2 smith:1 short:2 i1d:5 lx:1 klx:1 become:1 beta:1 viable:1 pairing:1 surprised:1 ik:3 fitting:1 headed:2 g4:1 acquired:1 forgetting:2 expected:4 behavior:1 p1:1 growing:1 td:7 automatically:1 little:1 estimating:2 underlying:1 null:1 what:1 contiguity:2 developed:2 finding:1 temporal:21 ti:1 exactly:1 grant:1 appear:2 before:3 tid:1 timing:1 consequence:1 sutton:5 encoding:1 oxford:1 solely:1 becoming:1 credited:1 black:1 plus:1 chose:2 dynamically:1 suggests:1 pavlov:1 limited:1 acknowledgment:1 testing:1 recursive:5 procedure:3 emanate:1 empirical:2 elicit:2 imminent:1 matching:1 integrating:3 get:1 cannot:3 barnet:4 judged:1 put:1 context:1 recency:1 deterministic:1 demonstrated:2 center:1 baum:2 missing:4 straightforward:1 iri:1 starting:1 duration:5 welch:2 formalized:1 immediately:4 importantly:1 his:2 classic:1 century:1 variation:1 krishnamurthy:1 aik:2 exact:4 soci:1 hypothesis:6 pa:1 recognition:1 observed:5 role:1 capture:1 environment:1 inaccessible:1 reward:1 trained:2 uations:1 xtixt:1 exposed:1 predictive:3 serve:1 learner:1 basis:1 preconditioning:3 po:1 joint:2 aii:1 represented:1 chapter:1 train:4 distinct:1 encoded:1 unconditioned:1 online:4 associative:5 sequence:5 propose:1 rescorla:3 product:1 combining:1 alxt:2 description:1 impaired:1 requirement:1 produce:4 help:1 measured:1 ij:5 strong:2 expt:11 involves:2 indicate:1 c:4 come:1 predicted:1 differ:1 posit:1 direction:1 correct:1 successor:2 require:1 generalization:3 elementary:1 extension:3 exploring:1 around:1 normal:1 cognition:1 predict:4 claim:1 estimation:5 integrates:1 label:3 pinette:2 currently:1 cole:4 footshock:6 successfully:2 reflects:1 weighted:1 impart:1 mit:1 inhibitor:1 gaussian:8 rather:3 occupied:1 cr:15 barto:3 mil:2 focus:1 pdfs:4 consistently:1 indicates:4 likelihood:4 lasted:2 baseline:2 helpful:1 inference:7 dayan:2 rigid:1 integrated:2 entire:1 hidden:15 relation:1 quasi:5 i1:2 ralph:1 among:3 animal:11 touretzk:1 constrained:2 integration:3 summed:1 represents:1 buzzer:2 future:2 cdhmm:5 stimulus:51 connectionist:1 employ:1 simultaneously:3 preserve:1 gamma:1 ime:1 national:1 phase:23 attempt:1 adjust:2 mixture:10 light:4 held:2 chain:15 implication:1 necessary:1 experience:3 old:1 initialized:2 minimal:1 psychological:1 modeling:1 earlier:1 ence:1 deviation:1 seventh:1 erlbaum:1 reported:3 varies:1 combined:1 st:5 density:6 thanks:1 sensitivity:1 probabilistic:1 off:2 lee:2 aix:1 together:2 continuously:1 cognitive:1 derivative:1 til:2 return:1 account:7 suggesting:1 ety:1 makov:1 coding:6 sec:8 coefficient:2 postulated:1 explicitly:1 onset:2 later:1 start:1 bayes:8 elicited:2 capability:1 il:1 oi:1 nathaniel:1 became:1 variance:4 characteristic:1 ufo:2 who:1 miller:12 ofthe:1 rabiner:1 matzel:1 weak:1 bayesian:3 produced:1 none:1 served:1 history:1 explain:2 simultaneous:6 failure:1 colleague:2 frequency:1 intertrial:1 obvious:1 associated:4 irvine:1 auditory:2 knowledge:2 formalize:1 agreed:1 reflecting:1 follow:1 response:6 anticipatory:1 furthermore:1 just:1 stage:1 governing:1 mode:3 indicated:1 effect:3 hypothesized:1 concept:2 true:1 inductive:1 moore:2 leibler:1 white:2 during:1 acquires:1 excitation:1 rat:2 pdf:2 complete:3 demonstrate:1 bring:1 meaning:1 suri:3 novel:1 recently:1 ji:1 conditioning:21 exponentially:2 association:6 occurred:1 he:5 interpret:1 mellon:1 appleton:1 dot:1 funded:2 similarity:1 depiction:1 inhibition:1 etc:1 multivariate:1 posterior:6 closest:1 compound:1 buffer:3 nonreinforcement:1 greater:1 preceding:2 converge:1 paradigm:5 maximize:2 monotonically:1 signal:1 ii:3 semi:1 multiple:2 desirable:1 ing:1 offer:2 long:3 concerning:1 serial:1 paired:2 qi:1 prediction:5 variant:3 basic:1 expectation:2 represent:5 robotics:1 fellowship:1 interval:1 source:1 pass:1 ufl:2 seem:1 effectiveness:1 call:1 presence:1 revealed:1 split:1 enough:2 canadian:1 iii:1 associating:1 click:4 vik:1 absent:1 expression:1 speech:2 gabriel:1 detailed:1 amount:1 repeating:1 reduced:1 supplied:1 tutorial:1 sign:1 disjoint:1 correctly:1 carnegie:1 discrete:1 hyperparameter:1 basal:1 key:1 threshold:1 changing:1 shock:1 backward:7 timestep:1 run:2 dst:1 arrive:1 family:1 reasonable:1 summarizes:1 capturing:1 followed:5 guaranteed:1 courville:2 annual:1 activity:1 strength:4 occur:1 simulate:1 pavlovian:2 department:1 conjugate:2 ur:1 wi:4 appealing:1 lp:1 restricted:1 pr:2 taken:3 computationally:1 equation:2 mutually:1 etworks:1 describing:1 count:1 mechanism:3 serf:1 gaussians:3 apply:1 progression:1 away:1 alternative:2 hat:2 existence:1 original:1 responding:1 dirichlet:1 remaining:1 trouble:1 assumes:1 iix:1 instant:1 exploit:1 restrictive:1 classical:8 society:1 quantity:1 occurs:4 rt:2 unable:1 separate:1 sci:1 length:1 modeled:3 relationship:2 index:1 acquire:3 unfortunately:1 holding:6 trace:1 implementation:1 proper:1 observation:26 markov:18 situation:1 extended:2 arbitrary:1 august:1 inferred:1 ordinate:1 david:1 california:1 emu:1 established:1 timeline:12 daw:1 able:1 suggested:1 bar:1 below:3 pattern:3 perception:1 oj:4 memory:15 royal:1 belief:3 unrealistic:1 event:2 natural:2 predicting:2 zr:1 recursion:1 representing:1 scheme:3 occupancy:4 review:1 literature:1 prior:4 hen:1 contributing:1 relative:2 bear:1 interesting:1 foundation:2 integrate:3 sufficient:1 consistent:1 editor:3 excitatory:1 last:1 aij:5 bias:1 allow:1 trr:3 institute:1 arnold:1 wagner:3 tolerance:1 dimension:2 world:5 transition:13 computes:2 sensory:3 forward:5 made:1 reinforcement:6 refinement:2 adaptive:8 schultz:3 simplified:1 collection:2 transaction:2 prokasy:1 approximate:4 implicitly:1 kullback:1 confirm:1 reproduces:2 vio:1 pittsburgh:1 assumed:1 continuous:4 latent:1 why:2 table:4 euroscience:1 learn:3 improving:1 protocol:1 domain:1 did:2 arrow:2 pgs:1 noise:2 hyperparameters:4 motivation:3 eural:1 tl:1 formalization:1 fails:1 experienced:1 exponential:1 tied:1 weighting:1 ix:2 croft:1 theorem:2 xt:7 symbol:1 offset:4 evidence:1 sit:1 intractable:1 merging:1 sequential:1 flashing:1 conditioned:8 illustrates:1 tio:1 gap:1 vii:1 led:1 likely:1 ganglion:1 forming:1 tch:2 unexpected:1 contained:1 nserc:1 scalar:1 reflex:1 determines:1 goal:1 presentation:4 absence:2 change:1 determined:1 total:1 pas:3 e:1 experimental:6 aaron:2 indicating:1 audio:1 phenomenon:3
1,126
2,027
TAP Gibbs Free Energy, Belief Propagation and Sparsity Lehel Csat?o and Manfred Opper Neural Computing Research Group School of Engineering and Applied Science Aston University, Birmingham B4 7ET, UK. [csatol,opperm]@aston.ac.uk Ole Winther Center for Biological Sequence Analysis, BioCentrum Technical University of Denmark, B208, 2800 Lyngby, Denmark. winther@cbs.dtu.dk Abstract The adaptive TAP Gibbs free energy for a general densely connected probabilistic model with quadratic interactions and arbritary single site constraints is derived. We show how a specific sequential minimization of the free energy leads to a generalization of Minka?s expectation propagation. Lastly, we derive a sparse representation version of the sequential algorithm. The usefulness of the approach is demonstrated on classification and density estimation with Gaussian processes and on an independent component analysis problem. 1 Introduction There is an increasing interest in methods for approximate inference in probabilistic (graphical) models. Such approximations may usually be grouped in three classes. In the first case we approximate self-consistency relations for marginal probabilities by a set of nonlinear equations. Mean field (MF) approximations and their advanced extensions belong to this group. However, it is not clear in general, how to solve these equations efficiently. This latter problem is of central concern to the second class, the Message passing algorithms, like Bayesian online approaches (for references, see e.g. [1]) and belief propagation (BP) which dynamically update approximations to conditional probabilities. Finally, approximations based on Free Energies allow us to derive marginal moments by minimising entropic loss measures. This method introduces new possibilities for algorithms and also gives approximations for the log-likelihood of observed data. The variational method is the most prominent member of this group. One can gain important insight into an approximation, when it can be derived by different approaches. Recently, the fixed points of the BP algorithm were identified as the stable minima of the Bethe Free Energy, an insight which led to improved approximation schemes [2]. While BP is good and efficient on sparse tree-like structures, one may look for an approxi- mation that works well in the opposite limit of densely connected graphs where individual dependencies are weak but their overall effect cannot be neglected. A interesting candidate is the adaptive TAP (ADATAP) approach introduced in [3] as a set of self-consistency relations. Recently, a message passing algorithm of Minka (termed expectation propagation) [1] was found to solve the ADATAP equations efficiently for models with Gaussian Process (GP) priors. The goal of this paper is three-fold. We will add a further derivation of ADATAP using an approximate free energy. A sequential algorithm for minimising the free energy generalises Minka?s result. Finally, we discuss how a sparse representation of ADATAP can be achieved for GP models, thereby extending previous sparse on-line approximations to the batch case [4]. We will specialize to probabilistic models on undirected graphs with nodes that are of the type    The set of   &"       !" $ #%  (1)  ?s encodes the dependencies between /01 32   the   random variables  (called likelihood in the , whereas the factorising term  ')(+*,*+*,(. following) usually encodes observed data at sites and also incorporates all local con straints of the (the range, discreteness, etc). Hence, depending on these constraints,    maybe discrete or continuous. Eq. (1) is a sufficiently rich and interesting class of models containing Boltzmann machines, models with Gaussian process priors [3], probabilistic independent component analysis [5] as well as Bayes belief networks and probabilistic neural networks (when the space of variables is augmented by auxiliary integration variables). 2 ADATAP approach from Gibbs Free Energy We use the minimization of an approximation to a Gibbs Free Energy derive the ADATAP approximation. 4 in order to re-  The Gibbs  Free Energy provides a method for computing marginal moments of as well as of 57698 within the same approach. It is defined by a constrained relative entropy minimization which is, for the present problem defined as  >=@? I H JLKM/0N A >: M0O+N A  (2) 8 BEDGF 5G698 AC ( ( <QP ( H M O N A where the brackets denote expectations with respect to the distribution and is M O N A H J TS@UVHW  \EA ] X9X9Y[YEZ Z . Finally, F 698 . shorthand for a vector with elements *  respect to its DR ( SinceH^ at _ the total minimum of 4 (with arguments) the minimizer in (2) is   =`? I: h  5i698 and the desired marginal just  , we conclude  M/0N M O Ngthat j >kElgm!8b=@ad? c ef4 I(g:V  8 adc e 4 < . moments of are (  n Lqsr  which is based on splitting 4 We will search for an approximation to 4 4po 4 , where 4to is the >u Gibbs free energy for a factorising model that is obtained from (1) by setting all  " attempts [6, 7] were based on a truncation of the power series r  . Previous expansion of 4 with respect to the &v at second order. While this truncation leads to the correct TAP equations for the large w limit of the so-called SK-model in statistical 4 ;: (< physics, its general significance is unclear. In fact, it will not be exact for a simple model with Gaussian likelihood. To make our approximation exact for such a case,;: we define  r (generalizing * $z;:  an{idea ;: of [8])  for an arbitrary Gaussian likelihood yx , 4 r x ;(: < z 4 5p4po . The main reason for this definition is the fact that 4 x  (< (g< (< is independent of the actual Gaussian likelihood x chosen to compute 4 ! This result depends crucially on the moment constraints in (2). Changes in a Gaussian likelihood can always be absorbed within the Lagrange-multipliers for the constraints. use this  \ 4po qGr We r 4 x , which universal form 4 x to define the ADATAP approximation as 4 by construction is exact for any Gaussian likelihood . Introducing appropriate Lagrange multipliers and , we get   r 4 x ;:  (<   >=Wk  c   57698   x     S U   x  (        n S U       o o o  o (     =Wk  5 4 o "! c  !$#  with (  q : R5 <     5 O'    O  q  O  5 O'  o , we have    o   o o  q : 68  (   R5  6 8   5 O (3)  ;y   !v   * Finally, setting  o&% *  o 5 < (4) 3 Sequential Algorithm ' \ ;:    The expression of 4 in terms of moments and Lagrange parameters \ ( o suggests that we may find local minima(< of 4  and o by iteratively alternating ( between updates of moments and Lagrange multipliers. Of special interest is the following sequential algorithm, which is a generalization of Minka?s EP [1] for Gaussian process classification to an arbritary model of the type eq. (1).      We choose a site and define the updates by using the saddle points of 4 with respect to the moments and Lagrange multipliers in the following sequential order (where is a diagonal matrix with elements  ): )+*, ,  fu . )78, c - c 9 4 , 4  fu:. )*, ,  u . )78! , c c- 9 ! 4, 4  fu:.   5  O / 0  ( 532  4 ' 6    o / 5    ) 5 9 , 4 ' 7<;,  / 5 - ,! 698 , 7< o ;  =/ 5  o 5 9 4 ' , .  The algorithm proceeds then by choosing a new site. The computation of T ( 5>2  4 ' can be performed efficiently using the Sherman-Woodbury formula because only one element   is changed in each update.   "/  10  ( 53 7, , 2785; 4 ' 6 v    o / 5  5 9 4 ,  / ) * , 698   o 7 , "/ 5   o! 5 9 , 4 78;, ( 3.1 Cavity interpretation ,   $?  B AX , @&Z   & & & & O   o 5 O' o At the fixed point, we may take  as the ADATAP approx    imation to the true marginal distribution of [3]. The sequential approach may thus be considered as a belief propagation algorithm for ADATAP. !   : Although  is usually not Gaussian, we can also derive the moments and from   the has a Gaussian distribution corresponding to x . This auxiliary Gaussian model < x   O  5 O '  q  and likelihood &x provides us also with an additional approxi    5  . This is useful when the coupling mation to the matrix of covariances via ' matrix must be adapted to a set of observations by maximum likelihood II. We will give an example of this for independent component analysis below. 8C    D  2  ( 32 4  It is important to understand the role of o and o within the ?cavity? approach to the  &v  , it is easy to show that  o M  N  and o TAP equations. Defining  M O N  M  N O   5 where the brackets denote an expectation with respect to the distribution of E HF E F E   E GF     C 2      c    &x  +           when node is all remaining variables x      deleted from the graph. This   statistics of corresponds to the empty ?cavity? at site . The  marginal distribution as computed by ADATAP is equivalent to the approximation that the cavity distribution is Gaussian. E   4 Examples 4.1 Models with Gaussian Process Priors , where the vector For  this class of models, we assume that the graph is embedded 0 in  is the restriction of a Gaussian process (random field) with  , to a set of            training inputs via   . is the posterior distribution corresponding to a local  likelihood model, when we set 5  '  and  the matrix  is obtained  from  a positive definite covariance kernel as D v D o  (  . The diagonal element D ' 9 is included in the likelihood term. 4 2 4 Our ADATAP approximation can be extended from the finite  set of inputs to the entire   space  by extending the auxiliary distribution x with its likelihoods M 0Gaussian N    &x  to a Gaussian process with mean and posterior covariance kernel D which (  A calculation similar to [4] leads to the approximates the posterior process. representation D  M  0 N     (       D o  (      q              D D o ( D o (  c D o ( (5) (6) D Algorithms for the update of ?s and ?s will usually suffer from time consuming matrix multiplications when w is large. This common problem for GP models can be overcome by a sparsity approximation which extends previous on-line approaches [4] to the batch ADATAP approach. The idea is to replace the current version x of the approximate Gaus sian with a further approximation  x for which both the the corresponding   as well as    are nonzero only, when the nodes  and  belong to a smaller subset  of nodes called ?basis vectors? (BV) of size  [4]. For fixed BV set, the parameters of  x are determined     !  and  by minimizing the relative entropy DGF  x x . This yields  Q ( with the #"Gw projection matrix  $&'% (' . Here  is the kernel matrix between  BVs and and  ' the kernel matrix between BVs and all nodes. The new distribution  x  a likelihood that contains  can be written in the form (1) with only BVs  D  4  x  $)% j       &$ %   5   +*   &$ %  + , O   *  ( ( (7) Eq. (7) can be used to compute the sparse approximation within the sequential algorithm. We will only give a brief discussion here. In order to recompute the appropriate ?cavity? parameters o and o when a new node is chosen by the algorithm, one removes a ?pseudo  $&%   variable? from the likelihood and recomputes of the remaining $&   $&%   the  statistics  % and the computation ones. When is in the BV set, then simply reduces to the previous one. We will demonstrate the significance of this approach for two examples.   4.2 Independent Component Analysis We consider  a measured signal -/. which is assumed to be an instantaneous linear mixing of sources corrupted with additive white Gaussian noise 0 that is, 21i q -/. . 03. (8) ( 1 where is a (time independent) mixing matrix and the noise vector is assumed to be without temporal correlations having time independent covariance matrix . We thus have the following likelihood for parameters and sources at time   4 ;  X  4 YZ  X  4 YIZ K1  j    4 ;  - . (9) . * W ( K 1 ( 0i 2 W K 1   and for all times . - . ( ( . . The aim  of independent com( ( the unknown ponent analysis is to recover quantities: the sources , the mixing matrix 1 and the noise covariance Wfrom the observed   2  W   data using the assumption of statistical independence . . . Following [5], we estimate the mix1 of the sources  an MLII procedure, i.e. by maximizing ing matrix and the noise covariance , by W K 1 j S`U W K 1 0yW  the Likelihood . The corresponding estimates are 4 ( (  1   M N ( M  N  M 1i  170 N '  . -. .  . . . and  These - 5 ' - 5 - again the structure of the* model estimates require averages over the posterior of which has eq. (1). They can be obtained efficiently using our sequential belief propagation M N M/ algorithm  N in an 1iterative EM fashion, where the E-step amounts to estimating and . . . with 1 W fixed and and the M-step consists of updating and . 5 Simulations 5.1 Classification with GPs This problem has been studied before [9, 4] using a sequential, sparse algorithm, based on a single sweep through the data only. Within the ADATAP approach we are able to perform multiple sweeps in order to achieve a self-consistent solution. The outputs are bi* 5 , and the likelihood is based on the probit model W ! K "bg $#jl&%0' j nary ! 'G "b/.0 0 S + U ( . ( ! and measures the noise level. The pre' O*) -,   5 O where   4 ;   * " l/o %, M  "bo  N # ! .1032 0 2 O 40 O q " b"  ! with , dictive distribution for a new test input is . . o to D eqs. ( (5). which is easily rewritten in terms of the parameters ?s and ?s according 5 5 We used the USPS dataset1 of with 6 7 u!u86 gray-scale handwritten digit images of size " training patterns and test patterns. For the kernel we choose the RBF kernel " "   :98;  " "  O . 0 ;O g >5  5=0 <; 5 < where is the dimension6 uEofu!u the inputs ( in D o ( 9 ; and are parameters. In the simulations we used random training this case), and examples. We performed simulations for different sizes of the BV set and compared multiple iterations with a single sweep through the dataset. The results are displayed in Fig. 1. > The lines show the average results of runs where the task was to classify the digits into fours/non-fours. Our results show that, in contrast to the online learning, the fluctuations caused by the order of presentation are diminished (marked with bars on the figure).      D     5.2 Density estimation with GPs ; Bayesian non-parametric models for density estimation can be defined [10] by parametris"K  A @ X X 2 2 ZZ B 2 ing densities ? as ? and using a Gaussian process prior over the space @ D" " of functions . Observing w data points C , we can express the predictive ' (,the *+*,*+GP ( - prior) as distribution (again, E denotes the expectation over ; S FGIH-J KMLON P Q R = S X FGIH-J TULWV FGIH Y J TUL\]N YZ[ S | 1 a}c e Q fh p h V R p X TUilGIHL~V P Q ^`_badc egfhh V R X TUijGIHLkV TUilGIH Y L \?? YZ[ Available from http://www.kernel-machines.org/data/ YZ[ TUilGIH Y L-monqpsrt ;vuxwy{z*w \ USPS: 4 <?> non 4 2.4 1 iterations 4 iterations 2.3 2.2 Test error % 2.1 2 1.9 1.8 1.7 1.6 1.5 50 150 250 350 450 550 #BV Figure 1: Results for classification for different BV sizes (x-axis) and multiple sweeps through the data. 4 3.5 3 2.5 2 1.5 1 0.5 0 0 0.2 0.4 0.6 0.8 1 Figure 2: The GP estimation (continuous line) of a mixture of Gaussians (dotted line) using u BVs.  ; 4 In the last expression, we have introduced an expectation X 2 ZB 2 a new, effective Gaussian  A @ over  obtained by multiplying the old prior and the term and normalizing by  . We assume that for sufficiently large w the integral over can be"performed K  M O by "bgLaplace?s N C , where method, leaving us with an approximate predictor of the form ? the brackets denote posterior expectation for a GP model with a kernel that is a solution to S U "  "  "     the integral equation D  ! ! 5 ! . The likelihood of D D D ; o ( ( "    o ( O  X ( Z the fields  * at the observation points is  . For any fixed  , we can apply the sparse ADATAP algorithm to this problem. After convergence of this inner loop, a new value of must be determined from (following a Laplace argument) -  M O "  N  until " global  convergence is achieved. To give a simplified toy example, we choose ! a kernel D which reproduces itself after convolution. Hence, the dependence is o ( we work and normalised at the end. We used a periodic kernel scaled out and with u for data in given by    ( D o D o " ( !  5     o " 5 !  q ?8  1  o " 5 has constant Fourier coefficients up to a cutoff frequency   4 ; g ! o ( C   o ,,@ ;    "  5 5 !  * in our simulations). For the experiment we are using artificial data from a mixture of two Gaussians (dotted line in Fig. 2). We apply the sparse algorithm with multiple sweeps through the data. The sparsity also avoids the numerical problems caused by a possible close to singular Gram matrix. For the experiments, the size of the BV set was not limited a priori, and a similar criterion as in [4] was chosen in order to decide whether a data u point should be included > uEu training data, only were retained in the BV set. in the BV set or not. As a result, for (continuous line in Fig. 2).  5.3 Independent Component Analysis We have tested the sequential algorithm on an ICA problem for local feature extraction in hand written 1 digits, i.e. extracting the different stroke styles [5] . We assumed positive components of (enforced by Lagrange multipliers) and a positive prior  W   `     . .  5  .    (10) As in [5] we used 500 handwritten ?3?s which are assumed to be generated by 25 hidden images. We compared a traditional parallel update algorithm with the sequential   belief propagation algorithm. Both algorithms have computational complexity w . We find that the sequential algorithm needs only on average 7 sweeps through the sites to reach the desired accuracy whereas the parallel one fails to reach the desired accuracy in 100 sweeps using a somewhat larger number of flops. The adaptive TAP method using the sequential belief propagation approach is also not more computationally expensive than the linear response method used in [5].   6 Conclusion and Outlook An obvious future direction for the ADATAP approach is the investigation of other minimization algorithms as an alternative to the EP approach outlined before. Also an extension of the sparse approximation to other non-GP models will be interesting. A highly important but difficult problem is the assessment of the accuracy of the approximation. Acknowledgments M. Opper is grateful to Lars Kai Hansen for suggesting the non-parametric density model. O. Winther thanks Pedro H?jen-S?rensen for the use of his Matlab code. The work is supported by EPSRC grant no. GR/M81601 and by the Danish Research Councils through Center for Biological Sequence Analysis. References [1] T.P. Minka. Expectation propagation for approximate Bayesian inference. PhD thesis, Dep. of Electrical Eng. and Comp. Sci.; MIT, 2000. [2] J. S. Yedidia, W. T. Freeman and Y. Weiss, Generalized Belief Propagation, to appear in Advances in Neural Information Processing Systems (NIPS?2000), MIT Press (2001). [3] M. Opper and O. Winther, Tractable approximations for probabilistic models: The adaptive TAP approach, Phys. Rev. Lett. 86, 3695 (2001). [4] L. Csat?o and M. Opper. Sparse Gaussian Processes. Neural Computation accepted (2001). [5] P.A.d.F.R. H?jen-S?rensen, O. Winther, and L. K. Hansen, Mean Field Approaches to Independent Component Analysis, Neural Computation accepted (2001). Available from http://www.cbs.dtu.dk/winther/ [6] T. Plefka, Convergence condition of the TAP equations for the infinite-ranged Ising spin glass model, J. Phys. A 15, 1971 (1982). [7] T. Tanaka, Mean-Field Theory of Boltzmann Machine Learning, Phys. Rev. E 58, 2302(1998). [8] G. Parisi and M. Potters, Mean-Field Equations for Spin Models with Orthogonal Interaction Matrices, J. Phys. A (Math. Gen.) 28, 5267 (1995). [9] L. Csat?o, E. Fokou?e, M. Opper, B. Schottky, and O. Winther. Efficient approaches to Gaussian process classification. In Advances in Neural Information Processing Systems, volume 12, (2000). [10] D.M. Schmidt. Continuous probability distributions from finite data. arXiv:physics/9808005 (1998)
2027 |@word version:2 simulation:4 crucially:1 covariance:6 eng:1 thereby:1 outlook:1 moment:8 series:1 contains:1 current:1 com:1 must:2 written:2 numerical:1 additive:1 lkv:1 remove:1 update:6 manfred:1 provides:2 recompute:1 node:6 math:1 org:1 specialize:1 shorthand:1 consists:1 ica:1 freeman:1 actual:1 increasing:1 estimating:1 adc:1 pseudo:1 temporal:1 scaled:1 uk:2 grant:1 appear:1 positive:3 before:2 engineering:1 local:4 limit:2 fluctuation:1 studied:1 dynamically:1 suggests:1 limited:1 range:1 bi:1 acknowledgment:1 woodbury:1 definite:1 digit:3 procedure:1 universal:1 projection:1 pre:1 get:1 cannot:1 close:1 restriction:1 equivalent:1 www:2 demonstrated:1 center:2 maximizing:1 splitting:1 insight:2 his:1 laplace:1 construction:1 exact:3 gps:2 element:4 expensive:1 updating:1 ising:1 observed:3 epsrc:1 ep:2 role:1 electrical:1 connected:2 complexity:1 neglected:1 grateful:1 predictive:1 basis:1 usps:2 po:2 easily:1 derivation:1 recomputes:1 effective:1 ole:1 artificial:1 ponent:1 choosing:1 larger:1 solve:2 kai:1 statistic:2 gp:7 itself:1 online:2 sequence:2 parisi:1 interaction:2 loop:1 gen:1 mixing:3 achieve:1 opperm:1 adatap:15 convergence:3 empty:1 extending:2 derive:4 depending:1 ac:2 coupling:1 measured:1 school:1 dep:1 eq:5 auxiliary:3 direction:1 correct:1 lars:1 require:1 generalization:2 tul:1 investigation:1 biological:2 extension:2 sufficiently:2 considered:1 bvs:4 entropic:1 estimation:4 birmingham:1 hansen:2 council:1 grouped:1 minimization:4 mit:2 gaussian:22 mation:2 always:1 aim:1 gaus:1 derived:2 ax:1 likelihood:18 contrast:1 glass:1 inference:2 lehel:1 entire:1 hidden:1 relation:2 overall:1 classification:5 priori:1 constrained:1 integration:1 special:1 marginal:6 field:6 having:1 extraction:1 r5:2 look:1 future:1 densely:2 individual:1 attempt:1 interest:2 message:2 possibility:1 highly:1 introduces:1 mixture:2 bracket:3 fu:3 integral:2 orthogonal:1 tree:1 old:1 re:1 desired:3 classify:1 introducing:1 subset:1 plefka:1 predictor:1 usefulness:1 gr:1 dependency:2 corrupted:1 periodic:1 thanks:1 density:5 winther:7 probabilistic:6 physic:2 again:2 central:1 thesis:1 containing:1 choose:3 m81601:1 dr:1 style:1 toy:1 suggesting:1 wk:2 coefficient:1 dictive:1 caused:2 ad:1 depends:1 performed:3 observing:1 bayes:1 hf:1 recover:1 parallel:2 spin:2 accuracy:3 efficiently:4 yield:1 weak:1 bayesian:3 handwritten:2 multiplying:1 comp:1 stroke:1 reach:2 phys:4 danish:1 definition:1 energy:11 frequency:1 minka:5 obvious:1 con:1 gain:1 dataset:1 ea:1 response:1 improved:1 wei:1 just:1 lastly:1 correlation:1 until:1 hand:1 nonlinear:1 assessment:1 propagation:10 gray:1 effect:1 ranged:1 multiplier:5 true:1 hence:2 alternating:1 iteratively:1 nonzero:1 white:1 gw:1 self:3 criterion:1 generalized:1 prominent:1 demonstrate:1 image:2 variational:1 instantaneous:1 recently:2 common:1 qp:1 b4:1 volume:1 jl:1 belong:2 interpretation:1 approximates:1 gibbs:6 approx:1 consistency:2 outlined:1 sherman:1 stable:1 etc:1 add:1 posterior:5 cbs:2 termed:1 minimum:3 additional:1 somewhat:1 signal:1 ii:1 multiple:4 reduces:1 ing:2 technical:1 generalises:1 calculation:1 minimising:2 expectation:8 arxiv:1 iteration:3 kernel:10 achieved:2 whereas:2 singular:1 source:4 leaving:1 nary:1 undirected:1 member:1 incorporates:1 extracting:1 easy:1 independence:1 identified:1 opposite:1 inner:1 idea:2 whether:1 expression:2 suffer:1 passing:2 matlab:1 useful:1 clear:1 maybe:1 amount:1 gih:3 http:2 rensen:2 dotted:2 csat:3 discrete:1 iz:1 express:1 group:3 four:2 deleted:1 discreteness:1 cutoff:1 schottky:1 graph:4 enforced:1 run:1 extends:1 decide:1 fold:1 quadratic:1 adapted:1 bv:9 constraint:4 bp:3 encodes:2 fourier:1 argument:2 according:1 smaller:1 em:1 rev:2 lyngby:1 equation:8 computationally:1 discus:1 imation:1 tractable:1 end:1 available:2 gaussians:2 rewritten:1 yedidia:1 apply:2 appropriate:2 batch:2 alternative:1 schmidt:1 denotes:1 remaining:2 straints:1 graphical:1 yx:1 k1:1 dgf:1 sweep:7 quantity:1 parametric:2 dependence:1 diagonal:2 traditional:1 unclear:1 sci:1 reason:1 denmark:2 potter:1 code:1 retained:1 minimizing:1 difficult:1 boltzmann:2 unknown:1 perform:1 observation:2 convolution:1 finite:2 t:1 displayed:1 defining:1 extended:1 flop:1 arbitrary:1 introduced:2 tap:8 tanaka:1 nip:1 able:1 bar:1 proceeds:1 usually:4 below:1 pattern:2 sparsity:3 belief:8 power:1 sian:1 advanced:1 scheme:1 aston:2 brief:1 dtu:2 lk:1 axis:1 gf:1 prior:7 multiplication:1 relative:2 embedded:1 loss:1 probit:1 interesting:3 consistent:1 factorising:2 changed:1 supported:1 last:1 free:11 truncation:2 allow:1 understand:1 normalised:1 sparse:10 overcome:1 opper:5 lett:1 gram:1 avoids:1 rich:1 dataset1:1 adaptive:4 mlii:1 simplified:1 approximate:6 cavity:5 approxi:2 global:1 reproduces:1 kmlon:1 conclude:1 assumed:4 consuming:1 continuous:4 search:1 iterative:1 sk:1 bethe:1 expansion:1 significance:2 main:1 noise:5 augmented:1 site:6 fig:3 fashion:1 fails:1 ueu:1 candidate:1 formula:1 specific:1 jen:2 badc:1 dk:2 concern:1 normalizing:1 sequential:14 phd:1 mf:1 entropy:2 generalizing:1 led:1 simply:1 saddle:1 absorbed:1 lagrange:6 bo:1 pedro:1 corresponds:1 minimizer:1 conditional:1 goal:1 presentation:1 marked:1 rbf:1 replace:1 change:1 included:2 determined:2 diminished:1 infinite:1 total:1 called:3 egf:1 accepted:2 latter:1 tested:1
1,127
2,028
Learning Discriminative Feature Transforms to Low Dimensions in Low Dimensions Kari Torkkola Motorola Labs, 7700 South River Parkway, MD ML28, Tempe AZ 85284, USA Kari.Torkkola@motorola.com http://members.home.net/torkkola Abstract The marriage of Renyi entropy with Parzen density estimation has been shown to be a viable tool in learning discriminative feature transforms. However, it suffers from computational complexity proportional to the square of the number of samples in the training data. This sets a practical limit to using large databases. We suggest immediate divorce of the two methods and remarriage of Renyi entropy with a semi-parametric density estimation method, such as a Gaussian Mixture Models (GMM). This allows all of the computation to take place in the low dimensional target space, and it reduces computational complexity proportional to square of the number of components in the mixtures. Furthermore, a convenient extension to Hidden Markov Models as commonly used in speech recognition becomes possible. 1 Introduction Feature selection or feature transforms are important aspects of any pattern recognition system. Optimal feature selection coupled with a particular classifier can be done by actually training and evaluating the classifier using all combinations of available features. Obviously this wrapper strategy does not allow learning feature transforms, because all possible transforms cannot be enumerated. Both feature selection and feature transforms can be learned by evaluating some criterion that reflects the ?importance? of a feature or a number of features jointly. This is called the filter configuration in feature selection. An optimal criterion for this purpose would naturally reflect the Bayes error rate. Approximations can be used, for example, based on Bhattacharyya bound or on an interclass divergence criterion. These are usually accompanied by a parametric estimation, such as Gaussian, of the densities at hand [6, 12]. The classical Linear Discriminant Analysis (LDA) assumes all classes to be Gaussian with a shared single covariance matrix [5]. Heteroscedastic Discriminant Analysis (HDA) extends this by allowing each of the classes have their own covariances [9]. Maximizing a particular criterion, the joint mutual information (MI) between the features and the class labels [1, 17, 16, 13], can be shown to minimize the lower bound of the classification error [3, 10, 15]. However, MI according to the popular definition of Shannon can be computationally expensive. Evaluation of the joint MI of a number of variables is plausible through histograms, but only for a few variables [17]. As a remedy, Principe et al showed in [4, 11, 10] that using Renyi?s entropy instead of Shannon?s, combined with Parzen density estimation, leads to expressions of mutual information with computational , where is the number of samples in the training set. This method complexity of can be formulated to express the mutual information between continuous variables and discrete class labels in order to learn dimension-reducing feature transforms, both linear   [15] and non-linear [14], for pattern recognition. One must note that regarding finding the extrema, both definitions of entropy are equivalent (see [7] pages 118,406, and [8] page 325). This formulation of MI evaluates the effect of each sample to every other sample in the transformed space through the Parzen density estimation kernel. This effect can also called as the ?information force?. Thus large/huge databases are hard to use due to the complexity.     To remedy this problem, and also to alleviate the difficulties in Parzen density estimation in high-dimensional spaces ( ), we present a formulation combining the mutual information criterion based on Renyi entropy with a semi-parametric density estimation method using Gaussian Mixture Models (GMM). In essence, Parzen density estimation is replaced by GMMs. In order to evaluate the MI, evaluating mutual interactions between mixture components of the GMMs suffices, instead of having to evaluate interactions between all pairs of samples. An approach that maps an output space GMM back to input space and again to output space through the adaptive feature transform is taken. This allows all of the computation to take place in the target low dimensional space. Computational complexity is reduced proportional to the square of the number of components in the mixtures. This paper is structured as follows. An introduction is given to the maximum mutual information (MMI) formulation for discriminative feature transforms using Renyi entropy and Parzen density estimation. We discuss different strategies to reduce its computational complexity, and we present a formulation based on GMMs. Empirical results are presented using a few well known databases, and we conclude by discussing a connection to Hidden Markov Models. 2 MMI for Discriminative Feature Transforms      " )( + ,. !-    0/ $1 # &% ' + 3 254  , * $6  9 7 , as samples of a continuous-valued random variable Given a set of training data , , and class labels as samples of a discrete-valued random variable , , the objective is to find a transformation (or its parameters ) to such that that maximizes , the mutual information (MI) between transformed data and class labels . The procedure is depicted in Fig. 1. To this end, we need to express as a function of the data set, , in a differentiable form. Once that is done, we can perform gradient ascent on as follows 7  89   7   +   7 F7 IH F 7 F + *;:=<?>@2A*;:CBED F * 25*:GBED F + F *  J >   * (1) To derive an expression for MI using a non-parametric density estimation method we apply Renyi?s quadratic entropy instead of Shannon?s entropy as described in [10, 15] because of its computational advantages. Estimating the density of as a sum of spherical Gaussians each centered at a sample , the expression of Renyi?s quadratic entropy of is K +  + N)O P'QSRUT6K  +    + L;M  9  2 2 2 IH  N)O P'Q   VR TEWX Y J HI I H N)O P'Q    Y Z J> J> 9 9 IH  + N + Y  ]  7   + N + Z  ]  7 _^`  + [ > Z J >\[  + Y N + Z ']  7   [ (2) Above, use is made of the fact that the convolution of two Gaussians is a Gaussian. Thus Renyi?s quadratic entropy can be computed as a sum of local interactions as defined by the kernel, over all pairs of samples. In order to use this convenient property, a measure of mutual information making use of quadratic functions of the densities would be desirable. Between a discrete variable and a continuous variable such a measure has been derived in [10, 15] as follows:  9 I I I 7  89  2 " RVT K   +    + B " RVT6K    K  +    + N " RVT K   +  K   K  +   + K Y  (3) We use for the number of samples in class ,  for  th sample regardless of its class,  and  for the same sample, but emphasizing that it belongs to class , with index  within the class. Expressing densities as their Parzen estimates with kernel width results in Z K ] HI I I 7   + $  2     Y   +  Y N +     ]  7  J> J> J>[  H I   I H I H  + Y +       Y  B     N ] 7 J >  [ J J > > I H  I I H  +  Z + Y    N ] 7 N     Z Y (4) J> J> J>[   Mutual information 7  +  can now be interpreted as an information potential induced samples of data in different classes. It is now straightforward to derive partial F 7  F + by which can accordingly be interpreted as an information force that other samples exert to sample + . The three components of the sum give rise to following three compo >  nents of the information force: Samples within the same class attract each other, All  samples regardless of class attract each other, and F  Samples F of different classes repel each other. This force, coupled with the latter factor +  * inside the sum in (1), tends to change the transform in such a way that the samples in transformed space  move  into the direction of the information force, and thus increase the MI criterion 7  +  . See [15] for details. Class labels: c g(w,x) Low dimensional features: y High-dimensional data: x Gradient Mutual Information I(c,y) (=Information potential) ?I ?w Figure 1: Learning feature transforms by maximizing the mutual information between class labels and transformed features. Each term in (4) consists of a double sum of Gaussians evaluated using the pairwise distance between the samples. The first component consists of a sum of these interactions within each class, the second of all interactions regardless of class, and the third of a sum of the interactions of each class against all other samples. The bulk of computation consists  Gaussians, and forming the sums of those. Information force, the of evaluating these gradient of , makes use of the same Gaussians, in addition to pairwise differences of the samples [15]. For large , complexity of is a problem. Thus, the rest of the paper explores possibilities of reducing the computation to make the method applicable to large databases. 7  '     3 How to Reduce Computation? In essence, we are trying to learn a transform that minimizes the class density overlap in the output space while trying to drive each class into a singularity. Since kernel density estimate results in a sum of kernels over samples, a divergence measure between the densities necessarily requires operations. The only alternatives to reduce this complexity are either to reduce , or to form simpler density estimates.     Two straightforward ways to achieve the former are clustering or random sampling. In this case clustering needs to be performed in the high-dimensional input space, which may be difficult and computationally expensive itself. A transform is then learned to find a representation that discriminates the cluster centers or the random samples belonging to different classes. Details of the densities may be lost, more so with random sampling, but at least this might bring the problem down to a computable level. The latter alternative can be accomplished by a GMM, for example. A GMM is learned in the low-dimensional output space for each class, and now, instead of comparing samples against each other, comparing samples against the components of the GMMs suffices. However, as the parameters of the transform are being learned iteratively, the will change at each iteration, and the GMMs need to be estimated again. There is no guarantee that the change to the transform and to the is so small that simple re-estimation based on previous GMMs would suffice. However, this depends on the optimization method used. +  + _ A further step in reducing computation is to compare GMMs of different classes in the output space against each other, instead of comparing the actual samples. In addition to the inconvenience of re-estimation, we lack now the notion of ?mapping?. Nothing is being transformed by from the input space to the output space, such that we could change the transform in order to increase the MI criterion. Although it would be possible now  to evaluate the effect of each sample to each mixture component, and the effect of each         , due to the double summing, we component to the MI, that is,   will pursue the mapping strategy outlined in the following section. 4 2 Y T T 4 Two GMM Mapping Strategies IO-mapping. If the GMM is available in the high-dimensional input space, those models can be directly mapped into the output space by the transform. Let us call this case the IO-mapping. Writing the density of class as a GMM with mixture components and  as their mixture weights we get Z K  I K     2 Z  Z   N  Z    Z  J> [ (5) We consider now only linear transforms. The transformed density in the low-dimensional output space is then simply  I K  +   2 Z  Z  + N  Z    Z  J> [  (6) Now, the mutual information in the output space between class labels and the densities as transformed GMMs can be expressed as a function of  , and it will be possible to evaluate to insert into (1). A great advantage of this strategy is that once the input space   GMMs have been created (by the EM-algorithm, for example), the actual training data needs not be touched at all during optimization! This is thus a very viable approach if the GMMs are already available in the high-dimensional input space (see Section 7), or if it is not too expensive or impossible to estimate them using the EM-algorithm. However, this might not be the case. F7 F OIO-mapping. An alternative is to construct a GMM model for the training data in the low-dimensional output space. Since getting there requires a transform, the GMM is constructed after having transformed the data using, for example, a random or an informed guess as the transform. Density estimated from the samples in the output space for class is K  I K  +   2 Z  Z  +  N Z Z J> [ (7) Once the output space GMM is constructed, the same samples are used to construct a GMM in the input space using the same exact assignments of samples to mixture components as the output space GMMs have. Running the EM-algorithm in the input space is now unnecessary since we know which samples belong to which mixture components. Similar strategy has been used to learn GMMs in high dimensional spaces [2]. Let us now use the notation of Eq.(5) to denote this density also in the input space. As a result, we have GMMs in both spaces and a transform mapping between the two.   The transform   can be learned as in the IO-mapping, by using the equalities  and   . This case will be called OIO-mapping. The biggest advantage is now avoiding to operate in the high-dimensional input space at all, not even the one time in the beginning of the procedure. Z 2  Z 2 Z Z 5 Learning the Transform through Mapped GMMs We present now the derivation of adaptation equations for a linear transform that apply to either mapping. The first step is to express the MI as a function of the GMM that is constructed in the output space. This GMM is a function of the transform matrix  , through the mapping of the input space GMM to the output space GMM. The second step is to compute its gradient   and to make use of it in the first half of Equation (1). F7 F 5.1 Information Potential as a Function of GMMs K  +   2  K +  K +  2 GMM in the output space for each class    is already expressed in (7). We need the following equalities: , where denotes the class prior, and    . H   J > K +   H ,     Let us denote the three terms in (3) as  , and N   . Then we have   H I I        _ ` ^       H 2 " RVT K  + + 2  RUT   XW +  N $   J> ! J > [ HI I  I  2     Z   Z   N  Z    B   Z  J > J > J > [ I Y 2 Y B  Y  N  (   % C      !      Y  "  "  C H I  H 2       )2 J>  C  2  Y   Y    C  2 Y   (8)  Y   2  [   2  To compact the notation, we change the indexing, and make the substitutions ,  ,   ,     ,   where , and is the total number of mixture components, and    . Now we can write  , , and in a convenient form.  [ [  H     H H H  I     I I   J > J > J >   2  IH Y IH      J> J>  (9) 5.2 Gradient of the Information Potential As each Gaussian mixture component is now a function of the corresponding input space component and the transform matrix  , it is straightforward (albeit tedious) to write the   . Since each of the three terms in gradient is composed of different sums of , we need its gradient as     F7 F  Y    Y   2 F F   Y   Y    (10)  [  [ [ where the input space GMM  parameters are  Y  2 Y N  and  Y  2  Y B   with the equalities Y  2  Y  and Y  2  Y   .      expresses the convolution of two mixture components in the output space. As we [ have those components in the high-dimensional input space, the gradient expresses also [ F F 7  C  2 F F  how this convolution in the output space changes, as  that maps the mixture components to the output space, is being changed. The mutual information measure is defined in terms of these convolutions, and maximizing it tends to find a  that (crudely stated) minimizes these convolutions between classes and maximizes them within classes. The desired gradient of the Gaussian with respect to the transform matrix is as follows: F     2 N    Y  >  7.N Y  Y  Y >   Y  B Y   Y  (11)  [ [ F F   The total gradient 7   can now be obtained simply by replacing C in (8) and (9) [ by the above gradient.   In evaluating 7 , the bulk of computation is inevaluating the  , the F componentwise F    . In addition, convolutions. Computational complexity is now the 7   requires F pairwise sums and differences of the mixture parameters in the input space, but these need only be computed once. 6 Empirical Results The first step in evaluating this approach is to compare its performance to the computationally more expensive MMI feature transforms that use Parzen density estimation. To this end, we repeated the pattern recognition experiments of [15] using exactly the same LVQ-classifier. These experiments were done using five publicly available databases that are very different in terms of the amount of data, dimension of data, and the number of training instances. For details of the data sets, please see [15]. OIO-mapping was used with 3-5 diagonal Gaussians per class to learn a dimension-reducing linear transform. Gradient ascent was used for optimization1. Results are presented in Tables 1 - 5. The last column denotes the original dimensionality of the data set. As a figure of the overall performance, the average over all five databases and all reduced dimensions, which ranged from one up to the original dimension minus one, was 69.6% for PCA, 77.8% for the MMI-Parzen combination, and 77.0% for the MMI-GMM combination (30 tests altogether). For LDA this figure cannot be calculated since some databases had a small and LDA can only produce features. The results are very satisfactory since the best we could hope for is performance equal to the MMI-Parzen combination. Thus a very significant reduction in computation caused only a minor drop in performance with this classifier.  "  " NA 7 Discussion We have presented a method to learn discriminative feature transforms using Maximum Mutual Information as the criterion. Formulating MI using Renyi entropy, and Gaussian 1 Example video clips can be viewed at http://members.home.net/torkkola/mmi. Table 1: Accuracy on the Phoneme test data set using LVQ classifier. Output dimenson PCA LDA MMI-Parzen MMI-GMM 1 7.6 5.1 15.5 21.4 2 70.0 66.0 68.5 70.4 3 76.8 74.7 75.2 76.8 4 81.1 80.2 80.2 80.2 6 84.2 82.8 82.6 82.6 9 87.3 86.0 85.3 87.7 20 90.0 - Table 2: Accuracy on the Landsat test data set using LVQ classifier. Output dimension PCA LDA MMI-Parzen MMI-GMM 1 41.2 42.5 65.1 65.0 2 81.5 75.7 82.0 80.4 3 85.8 86.2 86.4 86.1 4 87.8 87.2 86.2 88.3 9 89.4 88.8 87.6 87.4 15 90.3 90.0 89.5 89.1 36 90.4 - Table 3: Accuracy on the Letter test data set using LVQ classifier. Output dimension PCA LDA MMI-Parzen MMI-GMM 1 4.5 13.4 16.4 15.7 2 16.0 38.0 50.3 42.4 3 36.0 53.1 62.8 48.3 4 53.2 68.1 70.9 68.5 6 75.2 80.3 82.4 80.9 8 82.5 86.3 88.6 86.6 16 92.4 - Table 4: Accuracy on the Pipeline data set using LVQ classifier. Output dimension PCA LDA MMI-Parzen MMI-GMM 1 41.5 98.4 99.4 91.3 2 88.0 98.8 99.1 98.8 3 87.8 98.9 99.1 4 89.7 99.2 98.9 5 96.4 98.9 99.1 7 97.2 99.0 98.7 12 99.0 - Table 5: Accuracy on the Pima data set using LVQ classifier. Output dimension PCA LDA MMI-Parzen MMI-GMM 1 64.4 65.8 72.0 73.9 2 73.0 77.5 79.7 3 75.2 78.7 79.4 4 74.1 78.5 77.9 5 75.6 78.3 76.7 6 74.7 78.3 77.5 8 74.7 - Mixture Models as a semi-parametric density estimation method, allows all of the computation to take place in the low-dimensional transform space. Compared to previous formulation using Parzen density estimation, large databases become now a possibility. A convenient extension to Hidden Markov Models (HMM) as commonly used in speech recognition becomes also possible. Given an HMM-based speech recognition system, the state discrimination can be enhanced by learning a linear transform from some highdimensional collection of features to a convenient dimension. Existing HMMs can be converted to these high-dimensional features using so called single-pass retraining (compute all probabilities using current features, but do re-estimation using a the high-dimensional set of features). Now a state-discriminative transform to a lower dimension can be learned using the method presented in this paper. Another round of single-pass retraining then converts existing HMMs to new discriminative features. A further advantage of the method in speech recognition is that the state separation in the transformed output space is measured in terms of the separability of the data represented as Gaussian mixtures, not in terms of the data itself (actual samples). This should be advantageous regarding recognition accuracies since HMMs have the same exact structure. References [1] R. Battiti. Using mutual information for selecting features in supervised neural net learning. Neural Networks, 5(4):537?550, July 1994. [2] Sanjoy Dasgupta. Experiments with random projection. In Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence, pages 143?151, Stanford, CA, June30 - July 3 2000. [3] R.M. Fano. Transmission of Information: A Statistical theory of Communications. Wiley, New York, 1961. [4] J.W. Fisher III and J.C. Principe. A methodology for information theoretic feature extraction. In Proc. of IEEE World Congress On Computational Intelligence, pages 1712?1716, Anchorage, Alaska, May 4-9 1998. [5] K. Fukunaga. Introduction to statistical pattern recognition (2nd edition). Academic Press, New York, 1990. [6] Xuan Guorong, Chai Peiqi, and Wu Minhui. Bhattacharyya distance feature selection. In Proceedings of the 13th International Conference on Pattern Recognition, volume 2, pages 195 ? 199. IEEE, 25-29 Aug. 1996. [7] J.N. Kapur. Measures of information and their applications. Wiley, New Delhi, India, 1994. [8] J.N. Kapur and H.K. Kesavan. Entropy optimization principles with applications. Academic Press, San Diego, London, 1992. [9] Nagendra Kumar and Andreas G. Andreou. Heteroscedastic discriminant analysis and reduced rank HMMs for improved speech recognition. Speech Communication, 26(4):283?297, 1998. [10] J.C. Principe, J.W. Fisher III, and D. Xu. Information theoretic learning. In Simon Haykin, editor, Unsupervised Adaptive Filtering. Wiley, New York, NY, 2000. [11] J.C. Principe, D. Xu, and J.W. Fisher III. Pose estimation in SAR using an information-theoretic criterion. In Proc. SPIE98, 1998. [12] George Saon and Mukund Padmanabhan. Minimum bayes error feature selection for continuous speech recognition. In Todd K. Leen, Thomas G. Dietterich, and Volker Tresp, editors, Advances in Neural Information Processing Systems 13, pages 800? 806. MIT Press, 2001. [13] Janne Sinkkonen and Samuel Kaski. Clustering based on conditional distributions in an auxiliary space. Neural Computation, 14:217?239, 2002. [14] Kari Torkkola. Nonlinear feature transforms using maximum mutual information. In Proceedings of the IJCNN, pages 2756?2761, Washington DC, USA, July 15-19 2001. [15] Kari Torkkola and William Campbell. Mutual information in learning feature transformations. In Proceedings of the 17th International Conference on Machine Learning, pages 1015?1022, Stanford, CA, USA, June 29 - July 2 2000. [16] N. Vlassis, Y. Motomura, and B. Krose. Supervised dimension reduction of intrinsically low-dimensional data. Neural Computation, 14(1), January 2002. [17] H. Yang and J. Moody. Feature selection based on joint mutual information. In Proceedings of International ICSC Symposium on Advances in Intelligent Data Analysis, Rochester, New York, June 22-25 1999.
2028 |@word advantageous:1 retraining:2 nd:1 tedious:1 covariance:2 minus:1 reduction:2 wrapper:1 configuration:1 substitution:1 selecting:1 bhattacharyya:2 existing:2 current:1 com:1 comparing:3 must:1 drop:1 discrimination:1 half:1 intelligence:2 guess:1 accordingly:1 inconvenience:1 beginning:1 haykin:1 compo:1 simpler:1 five:2 anchorage:1 constructed:3 become:1 symposium:1 viable:2 consists:3 inside:1 pairwise:3 spherical:1 motorola:2 actual:3 becomes:2 estimating:1 notation:2 suffice:1 maximizes:2 interpreted:2 minimizes:2 pursue:1 informed:1 finding:1 extremum:1 transformation:2 guarantee:1 sinkkonen:1 every:1 exactly:1 classifier:9 local:1 todd:1 tends:2 limit:1 io:3 congress:1 tempe:1 might:2 exert:1 heteroscedastic:2 hmms:4 practical:1 rvt:3 lost:1 oio:3 procedure:2 empirical:2 convenient:5 projection:1 suggest:1 get:1 divorce:1 cannot:2 selection:7 impossible:1 writing:1 equivalent:1 map:2 center:1 maximizing:3 straightforward:3 regardless:3 notion:1 sar:1 target:2 enhanced:1 diego:1 exact:2 recognition:12 expensive:4 database:8 discriminates:1 complexity:9 joint:3 represented:1 kaski:1 derivation:1 london:1 artificial:1 stanford:2 plausible:1 valued:2 jointly:1 transform:21 itself:2 obviously:1 advantage:4 differentiable:1 net:3 interaction:6 adaptation:1 combining:1 achieve:1 az:1 getting:1 chai:1 double:2 cluster:1 transmission:1 produce:1 xuan:1 derive:2 pose:1 measured:1 minor:1 aug:1 eq:1 auxiliary:1 direction:1 filter:1 centered:1 suffices:2 alleviate:1 singularity:1 enumerated:1 extension:2 insert:1 marriage:1 great:1 mapping:12 nents:1 purpose:1 estimation:17 f7:4 proc:2 applicable:1 label:7 tool:1 reflects:1 hope:1 mit:1 gaussian:9 volker:1 derived:1 june:2 rank:1 attract:2 landsat:1 hidden:3 transformed:9 overall:1 classification:1 mutual:18 equal:1 once:4 construct:2 having:2 extraction:1 sampling:2 washington:1 unsupervised:1 intelligent:1 few:2 composed:1 divergence:2 replaced:1 william:1 huge:1 possibility:2 evaluation:1 mixture:17 partial:1 alaska:1 re:3 desired:1 instance:1 column:1 assignment:1 too:1 combined:1 density:26 explores:1 river:1 international:3 parzen:16 moody:1 na:1 again:2 reflect:1 potential:4 converted:1 accompanied:1 caused:1 depends:1 performed:1 icsc:1 lab:1 bayes:2 simon:1 rochester:1 minimize:1 square:3 publicly:1 accuracy:6 phoneme:1 mmi:17 drive:1 suffers:1 definition:2 evaluates:1 against:4 naturally:1 mi:12 popular:1 intrinsically:1 dimensionality:1 actually:1 back:1 campbell:1 supervised:2 methodology:1 improved:1 formulation:5 done:3 evaluated:1 leen:1 furthermore:1 crudely:1 hand:1 replacing:1 nonlinear:1 lack:1 lda:8 usa:3 effect:4 dietterich:1 ranged:1 remedy:2 former:1 equality:3 iteratively:1 satisfactory:1 round:1 during:1 width:1 please:1 essence:2 samuel:1 criterion:9 trying:2 theoretic:3 bring:1 volume:1 belong:1 expressing:1 significant:1 outlined:1 fano:1 had:1 own:1 showed:1 belongs:1 discussing:1 battiti:1 accomplished:1 minimum:1 george:1 july:4 semi:3 desirable:1 reduces:1 academic:2 motomura:1 hda:1 histogram:1 kernel:5 iteration:1 addition:3 rest:1 operate:1 ascent:2 south:1 induced:1 member:2 gmms:15 call:1 yang:1 iii:3 reduce:4 regarding:2 andreas:1 computable:1 expression:3 pca:6 speech:7 york:4 transforms:14 amount:1 clip:1 reduced:3 http:2 estimated:2 per:1 bulk:2 discrete:3 write:2 dasgupta:1 express:5 gmm:24 sum:11 convert:1 letter:1 uncertainty:1 place:3 extends:1 wu:1 separation:1 home:2 bound:2 hi:3 quadratic:4 ijcnn:1 aspect:1 fukunaga:1 formulating:1 kumar:1 structured:1 according:1 combination:4 saon:1 cbed:1 belonging:1 nagendra:1 em:3 separability:1 making:1 indexing:1 taken:1 pipeline:1 computationally:3 equation:2 discus:1 know:1 end:2 available:4 gaussians:6 operation:1 apply:2 alternative:3 altogether:1 original:2 thomas:1 assumes:1 clustering:3 running:1 denotes:2 xw:1 classical:1 objective:1 move:1 already:2 parametric:5 strategy:6 md:1 diagonal:1 gradient:12 distance:2 mapped:2 hmm:2 discriminant:3 index:1 difficult:1 pima:1 stated:1 rise:1 perform:1 allowing:1 convolution:6 markov:3 padmanabhan:1 kapur:2 january:1 immediate:1 vlassis:1 communication:2 dc:1 interclass:1 pair:2 repel:1 connection:1 componentwise:1 andreou:1 learned:6 delhi:1 rut:1 usually:1 pattern:5 video:1 overlap:1 difficulty:1 force:6 created:1 coupled:2 tresp:1 prior:1 proportional:3 filtering:1 principle:1 editor:2 changed:1 last:1 allow:1 india:1 dimension:14 calculated:1 evaluating:7 world:1 kari:4 commonly:2 adaptive:2 made:1 collection:1 san:1 compact:1 parkway:1 summing:1 conclude:1 unnecessary:1 discriminative:7 continuous:4 optimization1:1 table:6 learn:5 ca:2 necessarily:1 edition:1 nothing:1 repeated:1 xu:2 fig:1 biggest:1 ny:1 vr:1 wiley:3 third:1 renyi:9 touched:1 down:1 emphasizing:1 torkkola:6 mukund:1 ih:5 albeit:1 importance:1 entropy:12 depicted:1 simply:2 forming:1 expressed:2 conditional:1 viewed:1 formulated:1 lvq:6 shared:1 fisher:3 hard:1 change:6 reducing:4 called:4 total:2 pas:2 sanjoy:1 shannon:3 janne:1 principe:4 highdimensional:1 krose:1 latter:2 evaluate:4 avoiding:1
1,128
2,029
Hyperbolic Self-Organizing Maps for Semantic Navigation J?org Ontrup Neuroinformatics Group Faculty of Technology Bielefeld University D-33501 Bielefeld, Germany jontrup@techfak.uni-bielefeld.de Helge Ritter Neuroinformatics Group Faculty of Technology Bielefeld University D-33501 Bielefeld, Germany helge@techfak.uni-bielefeld.de Abstract We introduce a new type of Self-Organizing Map (SOM) to navigate in the Semantic Space of large text collections. We propose a ?hyperbolic SOM? (HSOM) based on a regular tesselation of the hyperbolic plane, which is a non-euclidean space characterized by constant negative gaussian curvature. The exponentially increasing size of a neighborhood around a point in hyperbolic space provides more freedom to map the complex information space arising from language into spatial relations. We describe experiments, showing that the HSOM can successfully be applied to text categorization tasks and yields results comparable to other state-of-the-art methods. 1 Introduction For many tasks of exploraty data analysis the Self-Organizing Maps (SOM), as introduced by Kohonen more than a decade ago, have become a widely used tool [1, 2]. So far, the overwhelming majority of SOM approaches have taken it for granted to use a flat space as their data model and, motivated by its convenience for visualization, have favored the (suitably discretized) euclidean plane as their chief ?canvas? for the generated mappings. However, even if our thinking is deeply entrenched with euclidean space, an obvious limiting factor is the rather restricted neighborhood that ?fits? around a point on a euclidean 2D surface. Hyperbolic spaces in contrast offer an interesting loophole. They are characterized by uniform negative curvature, resulting in a geometry such that the size of a neighborhood around a point increases exponentially with its radius . This exponential scaling behavior allows to create novel displays of large hierarchical structures that are particular accessible to visual inspection [3, 4]. Consequently, we suggest to use hyperbolic spaces also in conjunction with the SOM. The lattice structure of the resulting hyperbolic SOMs (HSOMs) is based on a tesselation of the hyperbolic space (in 2D or 3D) and the lattice neighborhood reflects the hyperbolic distance metric that is responsible for the non-intuitive properties of hyperbolic spaces. After a brief introduction to the construction of hyperbolic spaces we describe several computer experiments that indicate that the HSOM offers new interesting perspectives in the field of text-mining. 2 Hyperbolic Spaces Hyperbolic and spherical spaces are the only non-euclidean geometries that are homogeneous and have isotropic distance metrics [5, 6]. The geometry of H2 is a standard topic in Riemannian geometry (see, e.g. [7]), and the relationships for the area and the circumference of a circle of radius are given by              (1) These formulae exhibit the highly remarkable property that both quantities grow exponentially with the radius . It is this property that was observed in [3, 4] to make hyperbolic spaces extremely useful for accommodating hierarchical structures.   !  we must find suitable To use this potential for the SOM, we must solve two problems: discretization lattices on H2 to which we can ?attach? the SOM prototype vectors. after having constructed the SOM, we must somehow project the (hyperbolic!) lattice into ?flat space? in order to be able to inspect the generated maps. 2.1 Projections of Hyperbolic Spaces " To construct an isometric (i.e., distance preserving) embedding of the hyperbolic plane into a ?flat? space, we may use a Minkowski space [8]. In such a space, the squared distance between two points and is given by $# %& '&  #&(  % (  ' (  "  )$#+*,# (  .-  % * % (   /*  ' * ' (   (2) i.e., it ceases to be positive definite. Still, this is a space with zero curvature and its somewhat peculiar distance measure allows to construct an isometric embedding of the hyperbolic plane H2, given by #0 1 2 435 687 9%   :$2    :87 9'  365  ; 2  (3) $  2 7 where   are polar coordinates on the H2. Under this embedding, plane  )= - #  - the%  hyperbolic about the ' -axis. appears as the surface < swept out by rotating the curve ' From this embedding, we can construct two further ones, the so-called Klein model and the M Poincar?e model [5, 9] (we will use the latter to u visualize HSOMs below). Both achieve a proA jection of the infinite H2 into the unit disk, how1 B ever, at the price of distorting distances. The N Klein model is obtained by projecting the points C onto the plane along rays passing of through the origin (see Fig. 1). Obviously, D O 1 this projects all points of into the ?flat? unit disk of . (e.g., ). S The Poincar?e Model results if we add two further steps: first a perpendicular projection of Figure 1: Construction steps underlying the Klein Model onto the (?northern?) surface Klein and Poincar?e-models of the space H2 of the unit sphere centered at the origin (e.g., ), and then a stereographic projection of the ?northern? hemisphere onto the unit circle about the origin in the ground plane (point ). It turns out that the resulting projection of H2 has a number of pleasant properties, among them the preservation of < GHDF  ' >= ? #  - % A@ = B C <  ' JI K EDF G angles and the mapping of shortest paths onto circular arcs belonging to circles that intersect the unit disk at right angles. Distances in the original H2 are strongly distorted in its Poincar?e (and also in the Klein) image (cf. Eq. (5)), however, in a rather useful way: the mapping exhibits a strong ?fish-eye?-effect. The neighborhood of the H2 origin is mapped almost faithfully (up to a linear shrinkage factor of 2), while more distant regions become increasingly ?squeezed?. Since asymptotically the radial distances and the circumference grow both according to the same exponential law, the squeezing is ?conformal?, i.e., (sufficiently small) shapes painted onto H2 are not deformed, only their size shrinks with increasing distance from the origin. By translating the original H2, the fish-eye-fovea can be moved to any other part of H2, allowing to selectively zoom-in on interesting portions of a map painted on H2 while still keeping a coarser view of its surrounding context. 2.2 Tesselations of the Hyperbolic Plane To complete the set-up for a hyperbolic SOM we still need an equivalent of a regular grid in the hyperbolic plane. For the hyperbolic plane there exist an infinite number of tesselations with congruent polygons such that each grid point is surrounded by the same number of neighbors [9, 10]. Fig. 2 shows two example tesselations (for the minimal value of and for ), using the Poincar?e model for their visualization. While these tesselations appear non-uniform, this is only due to the fish-eye effect of the Poincar?e projection. In the original H2, each tesselation triangle has the same size.  )= I One way to generate these tesselations algorithmically is by repeated application of a suitable set of generators of their symmetry group to a (suitably sized, cf. below) ?starting triangle?, for more details cf. [11].   Figure 2: Regular triangle tesselations of the hyperbolic plane, projected into the unit disk using the Poincar?e mapping. The left tesselation shows the case where the minimal number ( ) of equilateral triangles meet at each vertex, the right figure was constructed with . In the Poincar?e projection, only sides passing through the origin appear straight, all other sides appear as circular arcs, although in the original space all triangles are congruent. 3 Hyperbolic SOM Algorithm We have now all ingredients required for a ?hyperbolic SOM?. We organize the nodes of a lattice as described above in ?rings? around an origin node. The numbers of nodes of such a lattice grows very rapidly (asymptotically exponentially) with the chosen lattice radius (its number of rings). For instance, a lattice with contains 1625 nodes. Each lattice node carries a prototype vector from some -dimensional feature space (if we wish to make any non-standard assumptions about the metric structure of this space, we would build this into the distance metric that is used for determining the best-match node). The SOM is then formed in the usual way, e.g., in on-line mode by C   C       B C K     in a radial           #+ *     (4)     :    *             with . However, since work on a hyperbolic lattice, "    weandnow   we have to determine both the neighborhood  the (squared) node distance " repeatedly determining the winner node and adjusting all nodes lattice neighborhood around according to the familiar rule  according to the natural metric that is inherited by the hyperbolic lattice. The simplest way to do this is to keep with each node a complex number to identify its position in the Poincar?e model. The node distance is then given (using the Poincar?e model, see e.g. [7]) as     "  arctanh  =  * *           (5)  The neighborhood can be defined as the subset of nodes within a certain graph distance (which is chosen as a small multiple of the neighborhood radius ) around . 4 Experiments Some introductory experiments where several examples illustrate the favorable properties of the HSOM as compared to the ?standard? euclidean SOM can be found in [11, 12]. A major example of the use of the SOM for text mining is the WEBSOM project [2]. 4.1 Text Categorization In order to apply the HSOM to natural text categorization, i.e. the assignment of natural language documents to a number of predefined categories, we follow the widely used vector-space-model of Information Retrieval (IR). For each document we construct a fea, where the components are determined by the frequency of which term ture vector occurs in that document. Following standard practice [13] we choose a term frequency inverse document frequency weighting scheme:  "    "         ! #"%$'&         (6)  "     where the term frequency   !  denotes the  number of times term  occurs in ") ( ,  the     number of documents in the training set and " the document frequency of  , i.e. the  number of documents occurs in.   The HSOM can be utilized for text categorization in the following manner. In a first step, the training set is used to adapt the weight vectors according to (4). During the second step, the training set is mapped onto the HSOM lattice. To this end, for each training example its best match node is determined such that "(      *    " (  +*      *     (7)  " (  -,   where  ")(  denotes the feature vector of document "( , as described above. After all examples have been presented to the net, each node is labelled with the union .  of all categories that belonged to the documents that were mapped to this node. A new, unknown text is then classified into the union . of categories which are associated with its winner node selected in the HSOM. Text Collection. We used the Reuters-215781 data set since it provides a well known baseline which is also used by other authors to evaluate their approaches, c.f. [14, 15]. We / 1 As compiled by David Lewis from the AT&T Research Lab in 1987. The data can be found at http://www.research.att.com/ lewis/ have used the ?ModApte? split, leading to 9603 training and 3299 test documents. After preprocessing, our training set contained 5561 distinct terms.    C    C  Performance Evaluation. The classification effectiveness is commonly measured in terms  of precision and recall [16], which can be estimated as         are the numbers of documents correctly classified, and      where  and  are the correctly not classified to category  , respectively. Analogous,  and  corresponding numbers of falsely classified documents.          For each node and each category  a confidence value is determined. It describes the number of training documents belonging to class  which were mapped to node . When retrieving documents from a given category  , we compare  for each node its associated against a threshold  . Documents from nodes with  become then included into the retrieval set. For nodes which contain a set of documents , the order of the , where . retrieval set is ranked by          " (   K   $   " (      K   In this way the number of retrieved documents can be controlled and we obtain the precision-recall-diagrams as shown in Fig. 3. In  order to compare the  HSOM?s performance for text categorization, we also evaluated a -nearest neighbor ( -NN) classifier with our training set. Apart from boosting methods [16] only  support vector machines [14] have shown better performances. The confidence level of a -NN classifier to assign document to class  is    " (   )(   $  " (  " (  (8)  ! #"%$'&  ( ( documents " for which   $ " (  "  is maximum. The assign- -NN   " (   I"   "(  where )( is the set of (  ment factor is 1,+if* belongs to category  and 0 otherwise. According to [14, 17] we & nearest neighbors. have chosen the C )  Text Categorization Results. The results of three experiments are shown -, in Table 1. We have compared a HSOM with rings and a tesselation with neighbors (summing up to 1306 nodes) to a  spherical standard euclidean SOM as described in [11] with approx. 1300 nodes, and  the -NN classifier. Our results indicate that the HSOM does not perform better than a -NN classifier, but to a certain extent also does not play significantly worse either. It is noticable that for less dominant categories the HSOM yields superior results to those of the standard SOM. This is due to the fact, that the nodes in H2 cover a much broader space and therefore offer more freedom to map smaller portions of the original dataspace with less distortions as compared to euclidean space.  As the -NN results suggest, other state-of-the-art techniques like support vector machines will probably lead to better numerical categorization results than the HSOM. However, since the main purpose of the HSOM is the visualization of relationships between texts and text categories, we believe that the observed categorization performance of the HSOM compares sufficiently well with the more specialized (non-visualization) techniques to warrant its efficient use for creating insightful maps of large bodies of document data. Table 1: Precision-recall breakeven points for the ten most prominent categories. SOM HSOM . -NN earn 90.0 90.2 93.8 acq 81.2 81.6 83.7 mny-fx 61.7 68.7 69.3 crude 70.3 78.8 84.7 grain 69.4 76.2 81.9 trade 48.8 56.8 61.9 interest 57.1 66.4 71.0 wheat 61.9 69.3 69.0 ship 54.8 61.8 77.5 corn 50.3 53.6 67.9 1 1 0.9 0.9 0.8 0.8 0.7 0.7 earn acq money?fx 0.6 0.6 0.4 0 0.2 0.4 0.6 0.8 0.4 1 0 0.2 . (a) -NN 0.4 0.6 0.8 1 (b) HSOM 2: 0.69 Figure 3: Precision-recall curves for the three most frequent categories earn, acq and money-fx. 4.2 Text Mining & Semantic Navigation A major advantage of the HSOM is its remarkable capability to map high-dimensional similarity relationships to a low-dimensional space which can be more easily handled and interpreted by the human observer. This feature and the particular ?fish-eye? capability motivates our approach to visualize whole text collections with the HSOM. It can be regarded as an interface capturing the semantic structure of a text database and provides a way to guide the users attention. In preliminary experiments we have labelled the nodes with glyphs corresponding to the categories of the documents mapped to that node. In Fig. 4 two HSOM views of the Reuters data set are shown. Note, that the major amount of data gets mapped to the outermost region, where the nodes of the HSOM make use of the large space offered by the hyperbolic geometry. During the unsupervised training process, the document?s categories were not presented to the HSOM. Nevertheless, several document clusters can be clearly identified. The two most prominent are the earn and acquisition region of the map, reflecting the large proportion of these categories in the Reuters-21578 collection. Note, that categories which are semantically similar are located beside each other, as can be seen in the corn, wheat, grain the interest, money-fx or the crude, ship area of the map. Additional to the category (glyph type) and the number of training documents per node (glyph size), the number of test documents mapped to each node is shown as the height of the symbol above the ground plane. In this way the HSOM can be used as a novelty detector in chronological document streams. For the Reuters-21578 dataset, a particular node strikes out. It corresponds to the small glyph tagged with the ?ship? label in Fig. 4. Only a few documents from the training collection are mapped to that node as shown by it?s relatively small glyph size. The large -value on the other hand indicates that it contains a large number of test documents, and is therefore probably semantically connected to a significant, novel event only contained in the test collection. The right image of Fig. 4 shows the same map, but the focal view now moved into the direction of the conspicious ?ship? node, resulting in a magnification of the corresponding area. A closer inspection reveals, that the vast majority (35 of 40) of the test documents describe an incident where an Iranian oil rig was attacked in the gulf. Although no document of the training set describes this incident (because the text collection is ordered by time and the attack took place ?after? the split into train and test set), the HSOM generalizes well and maps the semantic content of these documents to the proper area of the map, located between the regions for crude and ship.  The next example illustrates that the HSOM can provide more information about an unknown text than just it?s category. For this experiment we have taken movie reviews from the rec.art.movies.reviews newsgroup. Since all the reviews describe a certain movie, we retrieved their associated genres from the Internet Movie Database (http://www.imdb.com) to build a set of category labels for each document. The training set contained 8923 ran- money?fx ship trade corn wheat grain interest acq crude earn    Figure 4: The left figure shows a central view of the Reuters data. We used a HSOM with rings and a tesselation with neighbors. Ten different glyphs were used to visualize the ten most frequent categories. They were manually tagged to indicate the correspondence between category and symbol type. The glyph sizes and the -values (height above ground plane) reflect the number of training and test documents mapped to the corresponding node, respectively.  domly selected reviews (without their genre information) from films released before 2000. We then presented the system with five reviews from the film ?Atlantis?, a Disney cartoon released in 2001. The HSOM correctly classified all of the five texts as reviews for an animation movie. In Fig. 5 the projection of the five new documents onto the map with the previously acquired text collection is shown. It can be seen that there exist several clusters related to the animation genre. By moving the fovea of the HSOM we can now ?zoom? into that region which contains the five new texts. In the right of Fig. 5 it can be seen that all of the ?Atlantis? reviews where mapped to a node in immediate vicinity of documents describing other Disney animation movies. This example motivates the approach of ?semantic navigation? to rapidly visualize the linkage between unknown documents and previously acquired semantic concepts. Mulan Beauty and the beast Anastasia Pocahontas Hercules Aladin Atlantis Tarzan Chicken Run Dinosaur South Park Tarzan Mulan The Iron Giant Antz A Bug?s Life The Prince of Egypt      Figure 5: A HSOM with and a tesselation with neighbors was used to map movie rewies from newsgroup channels. In both figures, glyph size and -value indicate the number of texts related to the animation genre mapped to the corresponding node. Nodes exceeding a certain threshold were labelled with the title corresponding to the most frequently occuring movie mapped to that node. The underlined label in the right figure indicates the position of the node to which five new documents were mapped to. 5 Conclusion Efficient navigation in ?Sematic Space? requires to address two challenges: (i) how to create a low dimensional display of semantic relationship of documents, and (ii) how to obtain these relationships by automated text categorization. Our results show that the HSOM can provide a good solution to both demands simultaneously and within a single framework. The HSOM is able to exploit the peculiar geometric properties of hyperbolic space to successfully compress complex semantic relationships between text documents. Additionally, the use of hyperbolic lattice topology for the arrangement of the HSOM nodes offers new and attractive features for interactive ?semantic navigation?. Large document databases can be inspected at a glance while the HSOM provides additional information which was captured during a previous training step, allowing e.g. to rapidly visualize relationships between new documents and previously acquired collections. Future work will address more sophisticated visualization strategies based on the new approach, as well as the exploration of other text representations which might take advantage of hyperbolic space properties. References [1] T. Kohonen. Self-Organizing Maps. Springer Series in Information Sciences. 3rd edition, 2001. [2] Teuvo Kohonen, Samuel Kaski, Krista Lagus, Jarkko Saloj?arvi, Vesa Paatero, and Antti Saarela. Organization of a massive document collection. IEEE Transactions on Neural Networks, Special Issue on Neural Networks for Data Mining and Knowledge Discovery, 11(3):574?585, May 2000. [3] John Lamping and Ramana Rao. Laying out and visualizing large trees using a hyperbolic space. In Proceedings of UIST?94, pages 13?14, 1994. [4] T. Munzer. Exploring large graphs in 3D hyperbolic space. IEEE Computer Graphics and Applications, 18(4):18?23, July/August 1998. [5] H. S. M. Coxeter. Non Euclidean Geometry. Univ. of Toronto Press, Toronto, 1957. [6] J.A. Thorpe. Elementary Topics in Differential Geometry. Springer-Verlag, New York, 1979. [7] Frank Morgan. Riemannian Geometry: A Beginner?s Guide. Jones and Bartlett Publishers, Boston, London, 1993. [8] Charles W. Misner, J. A. Wheeler, and Kip S. Thorne. Gravitation. Freeman, 1973. [9] R. Fricke and F. Klein. Vorlesungen u? ber die Theorie der automorphen Funktionen, volume 1. Teubner, Leipzig, 1897. Reprinted by Johnson Reprint, New York, 1965. [10] W. Magnus. Noneuclidean Tesselations and Their Groups. Academic Press, 1974. [11] Helge Ritter. Self-organizing maps in non-euclidian spaces. In E. Oja and S. Kaski, editors, Kohonen Maps, pages 97?108. Amer Elsevier, 1999. [12] J. Ontrup and H. Ritter. Text categorization and semantic browsing with self-organizing maps on non-euclidean spaces. In Proc. of the PKDD-01, 2001. [13] G. Salton and C. Buckley. Term-weighting approaches in automatic text retrieval. Information Processing and Management, 24(5):513?523, 1988. [14] T. Joachims. Text categorization with support vector machines: learning with many relevant features. In Proc. of ECML-98, number 1398, pages 137?142, Chemnitz, DE, 1998. [15] Huma Lodhi, John Shawe-Taylor, Nello Cristianini, and Chris Watkins. Text classification using string kernels. In Todd K. Leen, Thomas G. Dietterich, and Volker Tresp, editors, Advances in Neural Information Processing Systems 13, pages 563?569. MIT Press, 2001. [16] F. Sebastiani, A. Sperduti, and N. Valdambrini. An improved boosting algorithm and its application to automated text categorization. In Proc. of CIKM-00, pages 78?85, 2000. [17] Y. Yang. An evaluation of statistical approaches to text categorization. Information Retrieval, 1-2(1):69?90, 1999.
2029 |@word deformed:1 faculty:2 proportion:1 suitably:2 disk:4 lodhi:1 euclidian:1 carry:1 contains:3 att:1 series:1 document:42 discretization:1 com:2 must:3 john:2 grain:3 numerical:1 distant:1 shape:1 leipzig:1 selected:2 tesselations:7 plane:13 inspection:2 isotropic:1 provides:4 boosting:2 node:39 toronto:2 attack:1 org:1 five:5 height:2 along:1 constructed:2 become:3 differential:1 retrieving:1 introductory:1 ray:1 manner:1 introduce:1 falsely:1 acquired:3 uist:1 behavior:1 pkdd:1 frequently:1 discretized:1 freeman:1 spherical:2 overwhelming:1 increasing:2 project:3 mulan:2 underlying:1 interpreted:1 string:1 giant:1 chronological:1 interactive:1 classifier:4 unit:6 appear:3 organize:1 atlantis:3 positive:1 before:1 todd:1 painted:2 meet:1 path:1 might:1 perpendicular:1 responsible:1 stereographic:1 practice:1 union:2 definite:1 poincar:10 wheeler:1 area:4 intersect:1 hyperbolic:34 significantly:1 projection:7 confidence:2 radial:2 regular:3 suggest:2 get:1 convenience:1 onto:7 context:1 www:2 equivalent:1 map:20 circumference:2 attention:1 starting:1 rule:1 regarded:1 embedding:4 coordinate:1 fx:5 analogous:1 limiting:1 techfak:2 construction:2 play:1 inspected:1 user:1 massive:1 homogeneous:1 origin:7 magnification:1 utilized:1 located:2 rec:1 coarser:1 database:3 observed:2 breakeven:1 region:5 wheat:3 connected:1 rig:1 trade:2 deeply:1 ran:1 cristianini:1 imdb:1 triangle:5 easily:1 polygon:1 kaski:2 genre:4 equilateral:1 surrounding:1 train:1 univ:1 distinct:1 describe:4 london:1 neighborhood:9 neuroinformatics:2 widely:2 solve:1 film:2 distortion:1 otherwise:1 obviously:1 advantage:2 net:1 took:1 propose:1 ment:1 fea:1 frequent:2 kohonen:4 relevant:1 rapidly:3 organizing:6 achieve:1 intuitive:1 moved:2 bug:1 cluster:2 sematic:1 congruent:2 categorization:13 ring:4 illustrate:1 measured:1 nearest:2 eq:1 strong:1 indicate:4 direction:1 radius:5 centered:1 human:1 exploration:1 translating:1 assign:2 preliminary:1 elementary:1 exploring:1 around:6 sufficiently:2 ground:3 magnus:1 mapping:4 visualize:5 major:3 released:2 purpose:1 domly:1 polar:1 favorable:1 proc:3 label:3 title:1 create:2 successfully:2 tool:1 reflects:1 faithfully:1 mit:1 clearly:1 gaussian:1 rather:2 shrinkage:1 beauty:1 volker:1 broader:1 conjunction:1 joachim:1 indicates:2 contrast:1 baseline:1 elsevier:1 nn:8 relation:1 germany:2 issue:1 among:1 classification:2 favored:1 spatial:1 art:3 special:1 noneuclidean:1 field:1 construct:4 having:1 cartoon:1 manually:1 park:1 jones:1 unsupervised:1 warrant:1 thinking:1 future:1 few:1 thorpe:1 oja:1 simultaneously:1 zoom:2 familiar:1 geometry:8 freedom:2 organization:1 interest:3 mining:4 highly:1 circular:2 evaluation:2 navigation:5 predefined:1 iranian:1 peculiar:2 closer:1 tree:1 euclidean:10 taylor:1 rotating:1 circle:3 prince:1 sperduti:1 minimal:2 instance:1 beginner:1 rao:1 cover:1 assignment:1 lattice:14 vertex:1 subset:1 uniform:2 teuvo:1 johnson:1 graphic:1 accessible:1 ritter:3 earn:5 squared:2 central:1 reflect:1 management:1 choose:1 worse:1 creating:1 leading:1 potential:1 de:3 jection:1 stream:1 vesa:1 teubner:1 view:4 observer:1 lab:1 portion:2 capability:2 inherited:1 acq:4 formed:1 ir:1 yield:2 identify:1 straight:1 ago:1 classified:5 detector:1 against:1 gravitation:1 acquisition:1 frequency:5 obvious:1 associated:3 riemannian:2 salton:1 dataset:1 adjusting:1 recall:4 knowledge:1 iron:1 sophisticated:1 reflecting:1 appears:1 isometric:2 follow:1 improved:1 amer:1 evaluated:1 shrink:1 strongly:1 leen:1 just:1 canvas:1 hand:1 glance:1 somehow:1 mode:1 grows:1 believe:1 glyph:8 effect:2 oil:1 contain:1 concept:1 dietterich:1 tagged:2 vicinity:1 semantic:11 attractive:1 visualizing:1 during:3 self:6 die:1 samuel:1 prominent:2 complete:1 occuring:1 egypt:1 interface:1 image:2 novel:2 charles:1 superior:1 specialized:1 ji:1 winner:2 exponentially:4 volume:1 significant:1 sebastiani:1 approx:1 rd:1 grid:2 focal:1 automatic:1 language:2 shawe:1 moving:1 similarity:1 surface:3 compiled:1 money:4 add:1 dominant:1 curvature:3 perspective:1 retrieved:2 hemisphere:1 apart:1 belongs:1 ship:6 certain:4 verlag:1 underlined:1 life:1 der:1 swept:1 preserving:1 seen:3 additional:2 somewhat:1 captured:1 morgan:1 determine:1 shortest:1 novelty:1 strike:1 july:1 preservation:1 ii:1 multiple:1 match:2 characterized:2 adapt:1 offer:4 sphere:1 retrieval:5 academic:1 controlled:1 metric:5 kernel:1 chicken:1 diagram:1 grow:2 publisher:1 probably:2 south:1 effectiveness:1 yang:1 split:2 ture:1 automated:2 fit:1 identified:1 topology:1 tarzan:2 prototype:2 reprinted:1 motivated:1 distorting:1 handled:1 bartlett:1 granted:1 linkage:1 gulf:1 passing:2 york:2 repeatedly:1 buckley:1 useful:2 pleasant:1 amount:1 ten:3 category:20 simplest:1 funktionen:1 generate:1 http:2 northern:2 exist:2 fish:4 estimated:1 arising:1 algorithmically:1 correctly:3 klein:6 per:1 cikm:1 group:4 threshold:2 nevertheless:1 vast:1 asymptotically:2 graph:2 run:1 angle:2 inverse:1 distorted:1 place:1 bielefeld:6 almost:1 scaling:1 comparable:1 capturing:1 internet:1 display:2 correspondence:1 flat:4 extremely:1 minkowski:1 corn:3 relatively:1 according:5 belonging:2 describes:2 smaller:1 increasingly:1 beast:1 helge:3 projecting:1 restricted:1 thorne:1 taken:2 visualization:5 previously:3 turn:1 describing:1 dataspace:1 conformal:1 end:1 generalizes:1 ramana:1 noticable:1 apply:1 hierarchical:2 original:5 compress:1 denotes:2 thomas:1 cf:3 exploit:1 build:2 arrangement:1 quantity:1 occurs:3 coxeter:1 strategy:1 usual:1 arctanh:1 anastasia:1 exhibit:2 fovea:2 distance:13 mapped:13 ontrup:2 majority:2 accommodating:1 modapte:1 topic:2 chris:1 extent:1 nello:1 laying:1 relationship:7 frank:1 theorie:1 negative:2 motivates:2 proper:1 unknown:3 perform:1 allowing:2 inspect:1 arc:2 attacked:1 ecml:1 immediate:1 ever:1 disney:2 august:1 introduced:1 david:1 required:1 kip:1 huma:1 address:2 able:2 below:2 belonged:1 challenge:1 loophole:1 suitable:2 event:1 natural:3 ranked:1 attach:1 scheme:1 movie:8 technology:2 brief:1 eye:4 axis:1 reprint:1 tresp:1 text:31 review:7 geometric:1 discovery:1 determining:2 law:1 beside:1 squeezed:1 interesting:3 squeezing:1 remarkable:2 ingredient:1 generator:1 h2:15 incident:2 offered:1 edf:1 editor:2 surrounded:1 dinosaur:1 keeping:1 antti:1 side:2 guide:2 ber:1 neighbor:6 outermost:1 curve:2 author:1 collection:10 commonly:1 projected:1 preprocessing:1 far:1 transaction:1 uni:2 keep:1 reveals:1 summing:1 decade:1 chief:1 table:2 additionally:1 channel:1 symmetry:1 complex:3 som:18 main:1 reuters:5 whole:1 animation:4 edition:1 repeated:1 body:1 chemnitz:1 fig:8 precision:4 position:2 wish:1 exceeding:1 exponential:2 crude:4 watkins:1 weighting:2 formula:1 navigate:1 showing:1 insightful:1 krista:1 symbol:2 cease:1 conspicious:1 illustrates:1 demand:1 browsing:1 boston:1 visual:1 contained:3 ordered:1 springer:2 corresponds:1 lewis:2 sized:1 consequently:1 labelled:3 price:1 content:1 included:1 infinite:2 determined:3 semantically:2 called:1 newsgroup:2 selectively:1 support:3 latter:1 evaluate:1 paatero:1
1,129
203
622 Atlas, Cole, Connor, EI-Sharkawi, Marks, Muthusamy and Barnard Performance Comparisons Between Backpropagation Networks and Classification Trees on Three Real-World Applications Ronald Cole Dept. of CS&E Oregon Graduate Institute Beaverton. Oregon 97006 Les Atlas Dept. of EE. Fr-10 University of Washington Seattle. Washington 98195 Jerome Connor, Mohamed EI-Sharkawi, and Robert J. Marks II University of Washington Etienne Barnard Carnegie-Mellon University Yeshwant Muthusamy Oregon Graduate Institute ABSTRACT Multi-layer perceptrons and trained classification trees are two very different techniques which have recently become popular. Given enough data and time, both methods are capable of performing arbitrary non-linear classification. We first consider the important differences between multi-layer perceptrons and classification trees and conclude that there is not enough theoretical basis for the clearcut superiority of one technique over the other. For this reason, we performed a number of empirical tests on three real-world problems in power system load forecasting, power system security prediction, and speaker-independent vowel identification. In all cases, even for piecewise-linear trees, the multi-layer perceptron performed as well as or better than the trained classification trees. Performance Comparisons 1 INTRODUCTION In this paper we compare regression and classification systems. A regression system can generate an output f for an input X, where both X and f are continuous and, perhaps, multi-dimensional. A classification system can generate an output class, C, for an input X, where X is continuous and multi-dimensional and C is a member of a finite alphabet. The statistical technique of Classification And Regression Trees (CART) was developed during the years 1973 (Meisel and Michalpoulos) through 1984 (Breiman el al). As we show in the next section, CART, like the multi-layer perceptron (MLP) , can be trained to solve the exclusive-OR problem. Furthermore, the solution it provides is extremely easy to interpret. Moreover, both CART and MLPs are able to provide arbitrary piecewise linear decision boundaries. Although there have been no links made between CART and biological neural networks, the possible applications and paradigms used for MLP and CART are very similar. The authors of this paper represent diverse interests in problems which have the commonality of being both important and potentially well-suited for trainable classifiers. The load forecasting problem, which is partially a regression problem, uses past load trends to predict the critical needs of future power generation. The power security problem uses the classifier as an interpolator of previously known states of the system. The vowel recognition problem is representative of the difficulties in automatic speech recognition due to variability across speakers and phonetic context. In each problem area, large amounts of real data were used for training and disjoint data sets were used for testing. We were careful to ensure that the experimental conditions were identical for the MLP and CART. We concentrated only on performance as measured in error on the test set and did not do any formal studies of training or testing time. (CART was, in general, quite a bit faster.) In all cases, even with various sizes of training sets, the multi-layer perceptron performed as well as or better than the trained classification trees. We also believe that integration of many of CART's well-designed attributes into MLP architectures could only improve the already promising performance of MLP's. 2 BACKGROUND 2.1 Multi-Layer Perceptrons The name "artificial neural networks" has in some commumbes become almost synonymous with MLP's trained by back-propagation. Our power studies made use of this standard algorithm (Rumelhart el ai, 1986) and our vowel studies made use of a conjugate gradient version (Barnard and Casasent, 1989) of back-propagation. In all cases the training data consisted of ordered pairs (X ,f)} for regression, or (X ,C)} for classification. The input to the network is X and the output is, after training, hopefully very close to f or C. When MLP's are used for regression, the output, f, can take on real values between 0 and 1. This normalized scale was used as the prediction value in the power forecasting problem. For MLP classifiers the output is formed by taking the (0,1) range of the output neurons and either thresholding or finding a peak. For example, in the vowel 623 624 Atlas, Cole, Connor, El-Sharkawi, Marks, Muthusamy and Barnard study we chose the maximum of the 12 output neurons to indicate the vowel class. 2.2 Classification and Regression Trees (CART) CART has already proven to be useful in diverse applications such as radar signal classification, medical diagnosis, and mass spectra classification (Breiman et ai, 1984). Given a set of training examples {(X ,C)}, a binary tree is constructed by sequentially partitioning the p -dimensional input space, which may consist of quantitative and/or qualitative data, into p -dimensional polygons. The trained classification tree divides the domain of the data into non-overlapping regions, each of which is assigned a class label C. For regression, the estimated function is piecewise constant over these regions. The first split of the data space is made to obtain the best global separation of the classes. The next step in CART is to consider the partitioned training examples as two completely unrelated sets-those examples on the left of the selected hyper-plane, and those on the right. CART then proceeds as in the first step, treating each subset of the training examples independently. A question which had long plagued the use of such sequential schemes was: when should the splitting stop? CART implements a novel, and very clever approach; splits continue until every training example is separated from every other, then a pruning criterion is used to sequentially remove less important splits. 2.3 Relative Expectations of MLP and CART The non-linearly separable exclusive-OR problem is an example of a problem which both MLP and CART can solve with zero error. The left side of Figure 1 shows a trained MLP solution to this problem and the right side shows the very simple trained CART solution. For the MLP the values along the arrows represent trained multiplicative weights and the values in the circles represent trained scalar offset values. For the CART figure, y and n represent yes or no answers to the trained thresholds and the values in the circles represent the output Y. It is interesting that CART did not train correctly for equal numbers of the four different input cases and that one extra example of one of the input cases was sufficient to break the symmetry and allow CART to train correctly. (Note the similarity to the well-known requirement of random and different initial weights for training the MLP). ~ y~ 08 Figure 1: The MLP and CART solutions to the exclusive-OR problem. Performance Comparisons CART trains on the exclusive-OR very easily since a piecewise-linear partition in the input space is a perfect solution. In general, the MLP will construct classification regions with smooth boundaries, whereas CART will construct regions with "sharp" comers (each region being, as described previously, an intersection of half planes). We would thus expect MLP to have an advantage when classification boundaries tend to be smooth and CART to have an advantage when they are sharper. Other important differences between MLP and CART include: For an MLP the number of hidden units can be selected to avoid overfitting or underfitting the data. CART fits the complexity by using an automatic pruning technique to adjust the size of the tree. The selection of the number of hidden units or the tree size was implemented in our experiments by using data from a second training set (independent of the first). An MLP becomes a classifier through an ad hoc application of thresholds or peak.picking to the output value(s). Great care has gone into the CART splitting rules while the usual MLP approach is rather arbitrary. A trained MLP represents an approximate solution to an optimization problem. The solution may depend on initial choice of weights and on the optimization technique used. For complex MLP's many of the units are independently and simultaneously adjusting their weights to best minimize output error. MLP is a distributed topology where a single point in the input space can have an effect across all units or analogously, one weight, acting alone, will have minimal affect on the outputs. CART is very different in that each split value can be mapped onto one segment in the input space. The behavior of CART makes it much more useful for data interpretation. A trained tree may be useful for understanding the structure of the data. The usefulness of MLP's for data interpretation is much less clear. The above points, when taken in combination, do not make a clear case for either MLP or CART to be superior for the best performance as a trained classifier. We thus believe that the empirical studies of the next sections, with their consistent performance trends, will indicate which of the comparative aspects are the most significant. 3 LOAD FORECASTING 3.1 The Problem The ability to predict electric power system loads from an hour to several days in the future can help a utility operator to efficiently schedule and utilize power generation. This ability to forecast loads can also provide information which can be used to strategically trade energy with other generating systems. In order for these forecasts to be useful to an operator, they must be accurate and computationally efficient. 3.2 Methods Hourly temperature and load data for the Seattle{facoma area were provided for us by the Puget Sound Power and Light Company. Since weekday forecasting is a more critical problem for the power industry than weekends, we selected the hourly data for 625 626 Atlas, Cole, Connor, El?Sharkawi, Marks, Muthusamy and Barnard all Tuesdays through Fridays in the interval of November 1, 1988 through January 31, 1989. These data consisted of 1368 hourly measurements that consisted of the 57 days of data collected. These data were presented to both the MLP and the CART classifier as a 6dimensional input with a single, real-valued output. The MLP required that all values be normalized to the range (0,1). These same normalized values were used with the CART technique. Our training and testing process consisted of training the classifiers on 53 days of the data and testing on the 4 days left over at the end of January 1989. Our training set consisted of 1272 hourly measurements and our test set contained 96 hourly readings. The MLP we used in these experiments had 6 inputs (Plus the trained constant bias term) 10 units in one hidden layer and one output. This topology was chosen by making use of data outside the training and test sets. 3.3 Results We used an 11 norm for the calculation of error rates and found that both techniques worked quite well. The average error rate for the :MLP was 1.39% and CART gave 2.86% error. While this difference (given the number of testing points) is not statistically significant. it is worth noting that the trained MLP offers performance which is at least as good as the current techniques used by the Puget Sound Power and Light Company and is currently being verified for application to future load prediction. 4 POWER SYSTEM SECURITY The assessment of security in a power system is an ongoing problem for the efficient and reliable generation of electric power. Static security addresses whether. after a disturbance. such as a line break or other rapid load change. the system will reach a steady state operating condition that does not violate any operating constraint and cause a "brown-out" or "black-out." The most efficient generation of power is achieved when the power system is operating near its insecurity boundary. In fact. the ideal case for efficiency would be full knowledge of the absolute boundaries of the secure regions. Due to the complexity of the power systems, this full knowledge is impossible. Load flow algorithms, which are based on iterative solutions of nonlinearly constrained equations, are conventionally used to slowly and accurately determine points of security or insecurity. In real systems the trajectories through the regions are not predictable in fine detail. Also these changes can happen too fast to compute new results from the accurate load flow equations. We thus propose to use the sparsely known solutions of the load flow equations as a training set The test set consists of points of unknown security. The error of the test set can then be computed by comparing the result of the trained classifier to load flow equation solutions. Our technique for converting this problem to a problem for a trainable classifier involves defining a training set ((X ,C?) where X is composed of real power, reactive power, and apparent power at another bus. This 3-dimensional input vector is paired with the corresponding security status (C=l for secure and C=O for insecure). Since Performance Comparisons the system was small, we were able to generate a large number of data points for training and testing. In fact, well over 20,000 total data points were available for the (disjoint) training and test sets. 4.1 Results We observed that for any choice of training data set size, the error rate for the MLP was always lower than the rate for the CART classifier. At 10,000 points of training data, the MLP had an error rate of 0.78% and CART has an error rate of 1.46%. While both of these results are impressive. the difference was statistically significant (p>.99). In order to gain insight into the reasons for differences in importance, we looked at classifier decisions for 2-dimensional slices of the input space. While the CART boundary sometimes was a better match, certain pathological difficulties made CART more error-prone than the MLP. Our other studies also showed that there were worse interpolation characteristics for CART. especially for sparse data. Apparently, starting with nonlinear combinations of inputs. which is what the MLP does. is better for the accurate fit than the stair-steps of CART. 5 SPEAKER-INDEPENDENT VOWEL CLASSIFICATION Speaker-independent classification of vowels excised from continuous speech is a most difficult task because of the many sources of variability that influence the physical realization of a given vowel. These sources of variability include the length of the speaker's vocal tract, phonetic context in which the vowel occurs, speech rate and syllable stress. To make the task even more difficult the classifiers were presented only with information from a single spectral slice. The spectral slice, represented by 64 DFf coefficients (0-4 kHz), was taken from the center of the vowel, where the effects of coarticulation with surrounding phonemes are least apparent. The training and test sets for the experiments consisted of featural descriptions, X, paired with an associated class, C. for each vowel sample. The 12 monophthongal vowels of English were used for the classes. as heard in the following words: beat. bit. bet, bat. roses. the, but, boot, book. bought, cot, bird. The vowels were excised from the wide variety of phonetic contexts in utterances of the TIMIT database, a standard acoustic phonetic corpus of continuous speech, displaying a wide range of American dialectical variation (Fisher et ai, 1986) (Lamel et ai, 1986). The training set consisted of 4104 vowels from 320 speakers. The test set consisted of 1644 vowels (137 occurrences of each vowel) from a different set of 100 speakers. The MLP consisted of 64 inputs (the DFf coefficients. each nonnalized between zero and one), a single hidden layer of 40 units, and 12 output units; one for each vowel category. The networks were trained using backpropagation with conjugate gradient optimization (Barnard and Casasent, 1989). The procedure for training and testing a network proceeded as follows: The network was trained on 100 iterations through the 4104 training vectors. The trained network was then evaluated on the training set and a different set of 1644 test vectors (the test set). The network was then trained for an additional 100 iterations and again evaluated on the training and test sets. This process was continued until the network had converged; convergence was observed as a 627 628 Atlas, Cole, Connor, EI?Sharkawi, Marks, Muthusamy and Barnard consistent decrease or leveling off of the classification percentage on the test data over successive sets of 100 iterations. The CART system was trained using two separate computer routines. One was the CART program from California Statistical Software; the other was a routine we designed ourselves. We produced our own routine to ensure a careful and independent test of the CART concepts described in (Breiman et ai, 1984). 5.1 Results In order to better understand the results, we performed listening experiments on a subset of the vowels used in these experiments. The vowels were excised from their sentence context and presented in isolation. Five listeners first received training in the task by classifying 900 vowel tokens and receiving feedback about the correct answer on each trial. During testing, each listener classified 600 vowels from the test set (50 from each category) without feedback. The average classification performance on the test set was 51%, compared to chance performance of 8.3%. Details of this experiment are presented in (Muthusamy et ai, 1990). When using the scaled spectral coefficients to train both techniques, the MLP correctly classified 47.4% of the test set while CART employing uni-variate splits performed at only 38.2%. One reason for the poor performance of CART with un i-variate splits may be that each coefficient (corresponding to energy in a narrow frequency band) contains little information when considered independently of the other coefficients. For example, reduced energy in the 1 kHz band may be difficult to detect if the energy in the 1.06 kHz band was increased by an appropriate amount. The CART classifier described above operates by making a series of inquiries about one frequency band at a time, an intuitively inappropriate approach. We achieved our best CART results, 46.4%, on the test set by making use of arbitrary hyper-planes (linear combinations) instead of univariate splits. This search-based approach gave results which were within 1% of the MLP results. 6 CONCLUSIONS In all cases the performance of the MLP was, in terms of percent error, better than CART. However, the difference in performance between the two classifiers was only significant (at the p >.99 level) for the power security problem. There are several possible reasons for the sometimes superior performance of the MLP technique, all of which we are currently investigating. One advantage may stem from the ability of MLP to easily find correlations between large numbers of variables. Although it is possible for CART to form arbitrary nonlinear decision boundaries, the efficiency of the recursive splitting process may be inferior to MLP's nonlinear fit. Another relative disadvantage of CART may be due to the successive nature of node growth. For example, if the first split that is made for a problem turns out, given the successive splits, to be suboptimal, it becomes very inefficient to change the first split to be more suitable. We feel that the careful statistics used in CART could also be advantageously applied to MLP. The superior performance of MLP is not yet indicative of best performance and it may turn out that careful application of statistics may allow further advance- Performance Comparisons ments in the MLP technique. It also may be possible that there would be input representations that would cause better performance for CART than for MLP. There have been new developments in trained statistical classifiers since the development of CART. More recent techniques, such as projection pursuit (Friedman and Stuetzle, 1984), may prove as good as or superior to MLP. This continued interplay between MLP techniques and advanced statistics is a key part of our ongoing research. Acknowledgements The authors wish to thank Professor R.D. Martin and Dr. Alan Lippman of the University of Washington Department of Statistics and Professors Aggoune, Damborg, and Hwang of the University of Washington Department of Electrical Engineering for their helpful discussions. David Cohn and Carlos Rivera assisted with many of the experiments. We also would like to thank Milan Casey Brace of Puget Power and Light for providing the load forecasting data. This work was supported by a National Science Foundation Presidential Young Investigator Award for L. Atlas and also by separate grants from the National Science Foundation and Washington Technology Center. References P. E. Barnard and D. Casasent, "Image Processing for Image Understanding with Neural Nets," Proc. Int. Joint Con! on Neural Nets, Washington, DC, June 18-22, 1989. L. Breiman, J.H. Friedman, R.A. Olshen, and CJ. Stone, Classification and Regression Trees, Wadsworth International, Belmont, CA, 1984. W. Fisher, G. Doddington, and K. Goudie-Marshall, "The DARPA Speech Recognition Research Database: Specification and Status," Proc. of the DARPA Speech Recognition Workshop, pp. 93-100, February 1986. J.H. Friedman and W. StuetzIe, "Projection Pursuit Regression," J. Amer. Stat. Assoc. 79, pp. 599-608, 1984. L. Lamel, R. Kassel, and S. Seneff, "Speech Database Development: Design and Analysis of the Acoustic-Phonetic Corpus," Proc. of the DARPA Speech Recognition Workshop, pp. 100-110, February 1986. W.S. Meisel and D.A. Michalpoulos, "A Partitioning Algorithm with Application in Pattern Classification and the Optimization of Decision Trees," IEEE Trans. Computers C-22, pp. 93-103. 1973. Y. Muthusamy. R. Cole, and M. Slaney. "Vowel Information in a Single Spectral Slice: Cochlcagrams Versus Spectrograms," Proc. ICASSP '90, April 3-6. 1990. (to appear) D.E. Rumelhart. G.E. Hinton, and RJ. Williams. "Learning Internal Representations by Error Propagation," Ch. 2 in Parallel Distributed Processing, D.E. Rumelhart, J.L. McClelland, and the PDP Research Group, MIT Press, Cambridge. MA, 1986. 629
203 |@word proceeded:1 trial:1 version:1 norm:1 weekday:1 rivera:1 initial:2 contains:1 series:1 dff:2 past:1 current:1 comparing:1 yet:1 must:1 belmont:1 ronald:1 happen:1 partition:1 remove:1 atlas:6 designed:2 treating:1 alone:1 half:1 selected:3 indicative:1 plane:3 provides:1 node:1 successive:3 five:1 along:1 constructed:1 become:2 qualitative:1 consists:1 prove:1 underfitting:1 leveling:1 cot:1 rapid:1 behavior:1 multi:8 company:2 little:1 inappropriate:1 becomes:2 provided:1 unrelated:1 moreover:1 mass:1 what:1 developed:1 finding:1 quantitative:1 every:2 growth:1 classifier:15 scaled:1 assoc:1 partitioning:2 unit:7 grant:1 medical:1 superiority:1 appear:1 hourly:5 engineering:1 interpolation:1 black:1 chose:1 plus:1 bird:1 graduate:2 range:3 gone:1 statistically:2 bat:1 testing:8 recursive:1 implement:1 backpropagation:2 lippman:1 procedure:1 stuetzle:1 area:2 excised:3 empirical:2 projection:2 word:1 vocal:1 onto:1 close:1 clever:1 selection:1 operator:2 context:4 impossible:1 influence:1 center:2 williams:1 starting:1 independently:3 splitting:3 rule:1 insight:1 continued:2 variation:1 feel:1 us:2 trend:2 rumelhart:3 recognition:5 sparsely:1 database:3 observed:2 electrical:1 region:7 trade:1 decrease:1 aggoune:1 rose:1 predictable:1 complexity:2 radar:1 trained:23 depend:1 segment:1 efficiency:2 basis:1 completely:1 comer:1 easily:2 joint:1 darpa:3 icassp:1 various:1 polygon:1 represented:1 listener:2 alphabet:1 train:4 separated:1 weekend:1 fast:1 surrounding:1 meisel:2 artificial:1 hyper:2 outside:1 quite:2 apparent:2 solve:2 valued:1 presidential:1 ability:3 statistic:4 hoc:1 advantage:3 interplay:1 net:2 propose:1 fr:1 realization:1 description:1 milan:1 seattle:2 convergence:1 requirement:1 comparative:1 perfect:1 generating:1 tract:1 help:1 stat:1 measured:1 received:1 implemented:1 c:1 involves:1 indicate:2 coarticulation:1 correct:1 attribute:1 biological:1 assisted:1 considered:1 plagued:1 great:1 predict:2 commonality:1 proc:4 label:1 currently:2 cole:6 mit:1 always:1 rather:1 avoid:1 dialectical:1 breiman:4 bet:1 casey:1 june:1 secure:2 detect:1 helpful:1 el:4 synonymous:1 hidden:4 classification:22 development:3 constrained:1 integration:1 wadsworth:1 equal:1 construct:2 washington:7 identical:1 represents:1 future:3 piecewise:4 strategically:1 pathological:1 composed:1 simultaneously:1 national:2 ourselves:1 vowel:22 friedman:3 mlp:47 interest:1 adjust:1 light:3 stair:1 accurate:3 capable:1 tree:15 divide:1 circle:2 theoretical:1 minimal:1 increased:1 industry:1 marshall:1 disadvantage:1 subset:2 usefulness:1 too:1 answer:2 peak:2 international:1 off:1 receiving:1 picking:1 analogously:1 again:1 slowly:1 dr:1 worse:1 slaney:1 book:1 american:1 inefficient:1 friday:1 puget:3 coefficient:5 int:1 oregon:3 ad:1 performed:5 multiplicative:1 break:2 apparently:1 carlos:1 parallel:1 timit:1 mlps:1 formed:1 minimize:1 phoneme:1 characteristic:1 efficiently:1 yes:1 identification:1 accurately:1 produced:1 trajectory:1 worth:1 insecurity:2 converged:1 classified:2 inquiry:1 reach:1 energy:4 frequency:2 mohamed:1 pp:4 associated:1 static:1 con:1 stop:1 gain:1 adjusting:1 popular:1 knowledge:2 cj:1 schedule:1 routine:3 back:2 day:4 april:1 amer:1 evaluated:2 furthermore:1 jerome:1 until:2 correlation:1 ei:3 cohn:1 nonlinear:3 hopefully:1 assessment:1 overlapping:1 propagation:3 perhaps:1 hwang:1 believe:2 name:1 effect:2 consisted:9 normalized:3 brown:1 concept:1 assigned:1 during:2 inferior:1 speaker:7 steady:1 criterion:1 stone:1 stress:1 lamel:2 temperature:1 yeshwant:1 percent:1 image:2 novel:1 recently:1 superior:4 physical:1 khz:3 interpretation:2 interpret:1 mellon:1 significant:4 measurement:2 connor:5 cambridge:1 ai:6 automatic:2 ongoing:2 had:4 specification:1 similarity:1 operating:3 impressive:1 own:1 showed:1 recent:1 phonetic:5 certain:1 binary:1 continue:1 seneff:1 additional:1 care:1 spectrogram:1 converting:1 determine:1 paradigm:1 signal:1 ii:1 violate:1 sound:2 rj:1 full:2 stem:1 smooth:2 alan:1 faster:1 match:1 calculation:1 offer:1 long:1 award:1 casasent:3 paired:2 prediction:3 regression:10 expectation:1 iteration:3 represent:5 sometimes:2 achieved:2 background:1 whereas:1 fine:1 interval:1 source:2 extra:1 brace:1 cart:51 tend:1 member:1 flow:4 bought:1 ee:1 near:1 noting:1 ideal:1 split:10 enough:2 muthusamy:7 easy:1 variety:1 affect:1 fit:3 gave:2 isolation:1 architecture:1 topology:2 variate:2 suboptimal:1 listening:1 whether:1 utility:1 forecasting:6 advantageously:1 speech:8 cause:2 useful:4 heard:1 clear:2 amount:2 band:4 concentrated:1 category:2 mcclelland:1 reduced:1 generate:3 percentage:1 estimated:1 disjoint:2 correctly:3 diverse:2 diagnosis:1 carnegie:1 group:1 key:1 four:1 threshold:2 interpolator:1 verified:1 utilize:1 year:1 almost:1 separation:1 decision:4 bit:2 layer:8 syllable:1 constraint:1 worked:1 software:1 aspect:1 extremely:1 performing:1 separable:1 martin:1 department:2 combination:3 poor:1 conjugate:2 across:2 partitioned:1 making:3 damborg:1 intuitively:1 taken:2 computationally:1 equation:4 previously:2 bus:1 turn:2 end:1 available:1 pursuit:2 spectral:4 appropriate:1 occurrence:1 ensure:2 include:2 beaverton:1 etienne:1 especially:1 february:2 kassel:1 already:2 question:1 looked:1 occurs:1 exclusive:4 usual:1 gradient:2 link:1 mapped:1 separate:2 thank:2 collected:1 reason:4 length:1 providing:1 difficult:3 olshen:1 robert:1 potentially:1 sharper:1 design:1 unknown:1 boot:1 neuron:2 finite:1 november:1 january:2 beat:1 defining:1 hinton:1 variability:3 nonnalized:1 dc:1 pdp:1 arbitrary:5 sharp:1 david:1 pair:1 required:1 nonlinearly:1 sentence:1 security:9 acoustic:2 california:1 narrow:1 hour:1 trans:1 address:1 able:2 proceeds:1 pattern:1 reading:1 program:1 reliable:1 power:22 critical:2 suitable:1 difficulty:2 disturbance:1 advanced:1 scheme:1 improve:1 technology:1 conventionally:1 utterance:1 featural:1 understanding:2 acknowledgement:1 relative:2 expect:1 generation:4 interesting:1 proven:1 versus:1 foundation:2 sufficient:1 consistent:2 thresholding:1 displaying:1 classifying:1 prone:1 token:1 supported:1 english:1 formal:1 side:2 allow:2 perceptron:3 institute:2 bias:1 wide:2 taking:1 understand:1 absolute:1 sparse:1 distributed:2 slice:4 boundary:7 feedback:2 world:2 author:2 made:6 employing:1 pruning:2 approximate:1 uni:1 status:2 global:1 sequentially:2 overfitting:1 investigating:1 corpus:2 conclude:1 spectrum:1 continuous:4 iterative:1 un:1 search:1 goudie:1 promising:1 nature:1 ca:1 symmetry:1 tuesday:1 complex:1 electric:2 domain:1 did:2 linearly:1 arrow:1 representative:1 wish:1 young:1 load:14 offset:1 ments:1 consist:1 workshop:2 sequential:1 importance:1 sharkawi:5 forecast:2 suited:1 intersection:1 univariate:1 ordered:1 contained:1 partially:1 scalar:1 ch:1 chance:1 ma:1 careful:4 barnard:8 fisher:2 professor:2 change:3 operates:1 acting:1 total:1 experimental:1 perceptrons:3 internal:1 mark:5 doddington:1 reactive:1 investigator:1 dept:2 trainable:2
1,130
2,030
Classifying Single Trial EEG: Towards Brain Computer Interfacing Benjamin Blankertz1?, Gabriel Curio2 and Klaus-Robert M?ller1,3 1 Fraunhofer-FIRST.IDA, Kekul?str. 7, 12489 Berlin, Germany 2 Neurophysics Group, Dept. of Neurology, Klinikum Benjamin Franklin, Freie Universit?t Berlin, Hindenburgdamm 30, 12203 Berlin, Germany 3 University of Potsdam, Am Neuen Palais 10, 14469 Potsdam, Germany       "!$#$&% ' ("')*,+-/.&!  0, 1"%02435 6  *%-/ 3 7  08# ) 943 Abstract Driven by the progress in the field of single-trial analysis of EEG, there is a growing interest in brain computer interfaces (BCIs), i.e., systems that enable human subjects to control a computer only by means of their brain signals. In a pseudo-online simulation our BCI detects upcoming finger movements in a natural keyboard typing condition and predicts their laterality. This can be done on average 100?230 ms before the respective key is actually pressed, i.e., long before the onset of EMG. Our approach is appealing for its short response time and high classification accuracy (>96%) in a binary decision where no human training is involved. We compare discriminative classifiers like Support Vector Machines (SVMs) and different variants of Fisher Discriminant that possess favorable regularization properties for dealing with high noise cases (inter-trial variablity). 1 Introduction The online analysis of single-trial electroencephalogram (EEG) measurements is a challenge for signal processing and machine learning. Once the high inter-trial variability (see Figure 1) of this complex multivariate signal can be reliably processed, the next logical step is to make use of the brain activities for real-time control of, e.g., a computer. In this work we study a pseudo-online evaluation of single-trial EEGs from voluntary self-paced finger movements and exploit the laterality of the left/right hand signal as one bit of information for later control. Features of our BCI approach are (a) no pre-selection for artifact trials, (b) state-of-the-art learning machines with inbuilt feature selection mechanisms (i.e., sparse Fisher Discriminant Analysis and SVMs) that lead to >96% classification accuracies, (c) non-trained users and (d) short response times. Although our setup was not tuned for speed, the a posteriori determined information transmission rate is 23 bits/min which makes our approach competitive to existing ones (e.g., [1, 2, 3, 4, 5, 6, 7]) that will be discussed in section 2. ? To whom correspondence should be addressed. Aims and physiological concept of BCI devices. Two key issues to start with when conceiving a BCI are (1) the definition of a behavioral context in which a subject?s brain signals will be monitored and used eventually as surrogate for a bodily, e.g., manual, input of computer commands, and (2) the choice of brain signals which are optimally capable to convey the subject?s intention to the computer. Concerning the behavioral context, typewriting on a computer keyboard is a highly overlearned motor competence. Accordingly, a natural first choice is a BCI-situation which induces the subject to arrive at a particular decision that is coupled to a predefined (learned) motor output. This approach is well known as a two alternative forced choice-reaction task (2AFC) where one out of two stimuli (visual, auditory or somatosensory) has to be detected, categorised and responded to by issuing one out of two alternative motor commands, e.g., pushing a button with either the left or right hand. A task variant without explicit sensory input is the voluntary, endogeneous generation of a ?go? command involving the deliberate choice between the two possible motor outputs at a self-paced rate. Here, we chose this latter approach so as to approximate the natural computer input situation of self-paced typewriting. Concerning the selection of brain signals related to such endogeneous motor commands we focussed here on one variant of slow brain potentials which are specifically related to the preparation and execution of a motor command, rather than reflecting merely unspecific modulations of vigilance or attention. Using multi-channel EEG-mapping it has been repeatedly demonstrated that several highly localised brain areas contribute to cerebral motor command processes. Specifically, a negative ?Bereitschaftspotential? (BP) precedes the voluntary initiation of the movement. A differential scalp potential distribution can be reliably demonstrated in a majority of experimental subjects with larger BP at lateral scalp positions (C3, C4) positioned over the left or right hemispherical primary motor cortex, respectively, consistenly correlating with the performing (right or left) hand [8, 9]. Because one potential BCI-application is with paralysed patients, one might consider to mimic the ?no-motor-output? of these individuals by having healthy experimental subjects to intend a movement but to withhold its execution (motor imagery). While it is true that brain potentials comparable to BP are associated with an imagination of hand movements, which indeed is consistent with the assumption that the primary motor cortex is active with motor imagery, actual motor performance significantly increased these potentials [10]. We therefore chose to instruct the experimental subjects to actually perform the typewriting finger movements, rather than to merely imagine their performance, for two reasons: first, this will increase the BP signal strength optimising the signal-to-noise ratio in BCI-related single trial analyses; and second, we propose that it is important for the subject?s task efficiency not to be engaged in an unnatural condition where, in addition to the preparation of a motor command, a second task, i.e., to ?veto? the very same movement, has to be executed. In the following section we will briefly review part of the impressive earlier research towards BCI devices (e.g., [1, 2, 3, 4, 5, 6, 7]) before experimental set-up and classification results are discussed in sections 3 and 4 respectively. Finally a brief conclusion in given. 2 A brief outline of BCI research Birbaumer et al. investigate slow cortical potentials (SCP) and how they can be selfregulated in a feedback scenario. In their thought translation device [2] patients learn to produce cortical negativity or positivity at a central scalp location at will, which is fed back to the user. After some training patients are able to transmit binary decisions in a 4 sec periodicity with accuracy levels up to 85% and therewith control a language support program or an internet browser. Pfurtscheller et al. built a BCI system based on event-related (de-)synchronisation (ERD/ERS, typically of the ? and central ? rhythm) for online classification of movement imaginations or preparations into 2?4 classes (e.g., left/right index finger, feet, tongue). Typical preprocessing techniques are adaptive autoregressive parameters, common spatial patterns (after band pass filtering) and band power in subject specific frequency bands. Classification is done by Fisher discriminant analysis, multi-layer neural networks or LVQ variants. In classification of exogeneous movement preparations, rates of 98%, 96% and 75% (for three subjects respectively) are obtained before movement onset 1 in a 3 classes task and trials of 8 sec [3]. Only selected, artifact free trials (less that 40%) were used. A tetraplegic patient controls his hand orthosis using the Graz BCI system. Wolpaw et al. study EEG-based cursor control [4], translating the power in subject specific frequency bands, or autoregressive parameters, from two spatially filtered scalp locations over sensorimotor cortex into vertical cursor movement. Users initially gain control by various kinds of motor imagery (the setting favours ?movement? vs. ?no movement? in contrast to ?left? vs. ?right?), which they report to use less and less as feedback training continues. In cursor control trials of at least 4 sec duration trained subjects reach accuracies of over 90%. Some subjects acquired also considerable control in a 2-d setup. 3 Acquisition and preprocessing of brain signals Experimental setup. The subject sat in a normal chair, relaxed arms resting on the table, fingers in the standard typing position at the computer keyboard. The task was to press with the index and little fingers the corresponding keys in a self-chosen order and timing (?self-paced key typing?). The experiment consisted of 3 sessions of 6 minutes each, preand postceeded by 60 seconds relaxing phase. All sessions were conducted on the same day with some minutes break inbetween. Typing of a total of 516 keystrokes was done at an average speed of 1 key every 2.1 seconds. Brain activity was measured with 27 Ag/AgCl electrodes at positions of the extended international 10-20 system, 21 mounted over motor and somatosensory cortex, 5 frontal and one occipital, referenced to nasion (sampled at 1000 Hz, band-pass filtered 0.05?200 Hz). Besides EEG we recorded an electromyogram (EMG) of the musculus flexor digitorum bilaterally (10?200 Hz) and a horizontal and vertical electrooculogram (EOG). In an event channel the timing of keystrokes was stored along with the EEG signal. All data were recorded with a NeuroScan device and converted to Matlab format for further analysis. The signals were downsampled to 100 Hz by picking every 10th sample. In a moderate rejection we sorted out only 3 out of 516 trials due to heavy measurement artifacts, while keeping trials that are contaminated by less serious artifacts or eye blinks. Note that 0.6% rejection rate is very low in contrast to most other BCI offline studies. The issue of preprocessing. Preprocessing the data can have a substantial effect on classification in terms of accuracy, effort and suitability of different algorithms. The question to what degree data should be preprocessed prior to classification is a trade-off between the danger of loosing information or overfitting and not having enough training samples for the classifier to generalize from high dimensional, noisy data. We have investigated two options: unprocessed data and preprocessing that was designed to focus on BP related to finger movement: (none) take 200 ms of raw data of all relevant channels; (<5 Hz) filter the signal low pass at 5 Hz, subsample it at 20 Hz and take 150 ms of all relevant channels (see Figure 1); Speaking of classification at a certain time point we strictly mean classification based on EEG signals until that very time point. The following procedure of calculating features of a single trial due to (<5 Hz) is easy applicable in an online scenario: Take the last 128 sample points of each channel (to the past relative from the given time point), apply a windowed (w(n) := 1 ? cos(n?/128)) FFT, keep only the coefficients corresponding to the pass 1 Precisely: before mean EMG onset time, for some trials this is before for others after EMG onset. 20 10 [?V] F3 F1 FZ F2 F4 CA5 CA3 CA1 CAZ CA2 CA4 CA6 C5 C3 C1 CZ C2 C4 C6 CP5 CP3 CP1 CPZ CP2 CP4 CP6 O1 average single trial feature 0 ?10 ?20 ?260 ?500 ?400 ?300 ?200 ?100 0 ?240 ?220 ?200 ?180 [ms] Figure 1: Averaged data and two single trials of right finger movements in channel C3. 3 values (marked by circles) of smoothed signals are taken as features in each channel. ?160 ?140 ?120 ? ?? Figure 2: Sparse Fisher Discriminant Analysis selected 68 features (shaded) from 405 input dimensions (27 channels ? 15 samples [150 ms]) of raw EEG data. band (bins 2?7, as bin 1 just contains DC information) and transform back. Downsampling to 20 Hz is done by calculating the mean of consecutive 5-tuple of data points. We investigated the alternatives of taking all 27 channels, or only the 21 located over motor and sensorimotor cortex. The 6 frontal and occipital channels are expected not to give strong contributions to the classification task. Hence a comparison shows, whether a classifier is disturbed by low information channels or if it even manages to extract information from them. Figure 1 depicts two single trial EEG signals at scalp location C3 for right finger movements. These two single trials are very well-shaped and were selected for resembling the the grand average over all 241 right finger movements, which is drawn as thick line. Usually the BP of a single trial is much more obscured by non task-related brain activity and noise. The goal of preprocessing is to reveal task-related components to a degree that they can be detected by a classifier. Figure 1 shows also the feature vectors due to preprocessing (<5 Hz) calculated from the depicted raw single trial signals. 4 From response-aligned to online classification We investigate some linear classification methods. Given a linear classifier (w, b) in separating hyperplane formulation (w> x+b = 0), the estimated label {1, ?1} of an input vector x ? N is y? = sign(w> x + b). If no a priori knowledge on the probability distribution of the data is available, a typical objective is to minimize a combination of empirical risk function and some regularization term that restrains the algorithm from overfitting to the training set {(xk , yk ) | k = 1, . . . , K}. Taking a soft margin loss function [11] yields the empirical risk function ?Kk=1 max(0, 1 ? yk (w> xk + b)). In most approaches of this type there is a hyper-parameter that determines the trade-off between risk and regularization, which has to be chosen by model selection on the training set2 . Fisher Discriminant (FD) is a well known classification method, in which a projection vector is determined to maximize the distance between the projected means of the two classes while minimizing the variance of the projected data within each class [13]. In the binary decision case FD is equivalent to a least squares regression to (properly scaled) class labels. Regularized Fisher Discriminant (RFD) can be obtained via a mathematical programming approach [14]: min 1/2 ||w||22 + C/K ||? ||22 w,b,? yk (w> xk + b) = 1 ? ?k 2 As subject to for k = 1, . . . , K this would be very time consuming in k-fold crossvalidation, we proceed similarly to [12]. filter <5 Hz <5 Hz none none ch?s mc all mc all FD 3.7?2.6 3.3?2.5 18.1?4.8 29.3?6.1 RFD 3.3?2.2 3.1?2.5 7.0?4.1 7.5?3.8 SFD 3.3?2.2 3.4?2.7 6.4?3.4 7.0?3.9 SVM 3.2?2.5 3.6?2.5 8.5?4.3 9.8?4.4 k-NN 21.6?4.9 23.1?5.8 29.6?5.9 32.2?6.8 Table 3: Test set error (? std) for classification at 120 ms before keystroke; ?mc? refers to the 21 channels over (sensori) motor cortex, ?all? refers to all 27 channels. where ||?||2 denotes the `2 -norm (||w||22 = w> w) and C is a model parameter. The constraint yk (w> xk + b) = 1 ? ?k ensures that the class means are projected to the corresponding class labels, i.e., 1 and ?1. Minimizing the length of w maximizes the margin between the projected class means relative to the intra class variance. This formalization above gives the opportunity to consider some interesting variants, e.g., Sparse Fisher Discriminant (SFD) uses the `1 -norm (||w||1 = ?|wn |) on the regularizer, i.e., the goal function is 1/N ||w||1 + C/K ||? ||22 . This choice favours solutions with sparse vectors w, so that this method also yields some feature selection (in input space). When applied to our raw EEG signals SFD selects 68 out of 405 input dimensions that allow for a left vs. right classification with good generalization. The choice coincides in general with what we would expect from neurophysiology, i.e., high loadings for electrodes close to left and right hemisphere motor cortices which increase prior to the keystroke, cf. Figure 2. But here the selection is automatically adapted to subject, electrode placement, etc. Our implementation of RFD and SFD uses the cplex optimizer [15]. Support Vector Machines (SVMs) are well known for their use with kernels [16, 17]. Here we only consider linear SVMs: min 1/2 ||w||22 + C/K ||? ||1 w,b,? yk (w> xk + b) subject to 1 ? ?k , and ?k 0 The choice of regulization keeps a bound on the Vapnik-Chervonenkis dimension small. In an equivalent formulation the objective is to maximize the margin between the two classes (while minimizing the soft margin loss function)3. For comparision we also employed a standard classifier of different type: k-Nearest-Neighbor (k-NN) maps an input vector to that class to which the majority of the k nearest training samples belong. Those neighbors are determined by euclidean distance of the corresponding feature vectors. The value of k chosen by model selection was around 15 for processed and around 25 for unprocessed data. Classification of response-aligned windows. In the first phase we make full use of the information that we have regarding the timing of the keystrokes. For each single trial we calculate a feature vector as described above with respect to a fixed timing relative to the key trigger (?response-aligned?). Table 3 reports the mean error on test sets in a 10?10fold crossvalidation for classifying in ?left? and ?right? at 120 ms prior to keypress. Figure 4 shows the time course of the classification error. For comparison, the result of EMG-based classification is also displayed. It is more or less at chance level up to 120 ms before the keystroke. After that the error rate decreases rapidly. Based upon this observation we chose t =?120 ms for investigating EEG-based classification. From Table 3 we see that FD works well with the preprocssed data, but as dimensionality increases the performance breaks down. k-NN is not successful at all. The reason for the failure is that the variance in the discriminating directions is much smaller that the variance in other directions. So using the euclidean metric is not an appropirate similarity measure for this purpose. All regularized discriminative classifiers attain comparable results. For preprocessed data a very low 3 We used the implementation LIBSVM by Chang and Lin, available along with other implementations from        ! . 60 12 classification error [%] ?120 ms? 50 10 40 8 30 6 20 4 10 0 ?1000 2 EMG EEG ?800 ?600 ?400 ?200 0 [ms] 0 ?200 ?100 0 [ms] Figure 4: Comparison of EEG (<5 Hz, mc, SFD) and EMG based classification with respect to the endpoint of the classification interval. The right panel gives a details view: -230 to 50 ms. error rate between 3% and 4% can be reached without a significant difference between the competing methods. For the classification of raw data the error is roughly twice as high. The concept of seeking sparse solution vectors allows SFD to cope very well with the high dimensional raw data. Even though the error is twice as high compared to the the minimum error, this result is very interesting, because it does not rely on preprocessing. So the SFD approach may be highly useful for online situations, when no precursory experiments are available for tuning the preprocessing. The comparison of EEG- and EMG-based classification in Figure 4 demonstrates the rapid response capability of our system: 230 ms before the actual keystroke the classification rate exceeds 90%. To assess this result it has to be recalled that movements were performed spontaneously. At ?120 ms, when the EMG derived classifier is still close to chance, EEG based classification becomes already very stable with less than 3.5% errors. Interpreting the last result in the sense of a 2AFC gives an information transfer rate of 60/2.1B ? 22.9 [bits/min], where B = log2 N + p log2 p+(1? p) log2 (1?p/N?1) is the number of bits per selection from N = 2 choices with success probability p = 1 ? 0.035 (under some uniformity assumptions). Classification in sliding windows. The second phase is an important step towards online classification of endogeneous brain signals. We have to refrain from using event timing information (e.g., of keystrokes) in the test set. Accordingly, classification has to be performed in sliding windows and the classifier does not know in what time relation the given signals are to the event?maybe there is even no event. Technically classification could be done as before, as the trained classifiers can be applied to the feature vectors calculated from some arbitrary window. But in practice this is very likely to lead to unreliable results since those classifiers are highly specialized to signals that have a certain time relation to the response. The behavior of the classifier elsewhere is uncertain. The typical way to make classification more robust to time shifted signals is jittered training. In our case we used 4 windows for each trial, ending at -240, -160, -80 and 0 ms relative to the response (i.e., we get four feature vectors from each trial). Movement detection and pseudo-online classification. Detecting upcoming events is a crucial point in online analysis of brain signals in an unforced condition. To accomplish this, we employ a second classifier that distinguishes movement events from the ?rest?. Figures 5 and 6 display the continuous classifier output w> x + b (henceforth called graded) for left/right and movement/rest distinction, respectively. For Figure 5, a classifier was trained as described above and subsequently applied to windows sliding over unseen test samples yielding ?traces? of graded classifier outputs. After doing this for several train/test set splits, the borders of the shaded tubes are calculated as 5 and 95 percentile values of          1      0.5       0      ?0.5       ?1    ?1.5    1.5 ?1000                                                 ?750                         ?500                                                 ?250                         0                            250                                                       500                               750     right    left                    [ms] Figure 5: Graded classifier output for left/right distinctions.       1      0    ?1       ?2 2 ?1000                                           ?750                      ?500                                           ?250                      0                      250     median              90 percentile              10,     5?95    perc.   tube                                                         500 750 [ms] Figure 6: Graded classifier output for movement detection in endogenous brain signals. those traces, thin lines are at 10 and 90 percent, and the thick line indicates the median. At t =?100 ms the median for right events in Figure 5 is approximately 0.9, i.e., applying the classifier to right events from the test set yielded in 50% of the cases an output greater 0.9 (and in 50% an output less than 0.9). The corresponding 10-percentile line is at 0.25 which means that the output to 90% of the right events was greater than 0.25. The second classifier (Figure 6) was trained for class ?movement? on all trials with jitters as described above and for class ?rest? in multiple windows between the keystrokes. The preprocessing and classification procedure was the same as for left vs. right. The classifier in Figure 5 shows a pronounced separation during the movement (preparation and execution) period. In other regions there is an overlap or even crossover of the classifier outputs of the different classes. From Figure 5 we observe that the left/right classifier alone does not distinguish reliably between ?movement? and ?no movement? by the magnitude of its output, which explains the need for a movement detector. The elevation for the left class is a little less pronounced (e.g., the median is ?1 at t =0 ms compared to 1.2 for right events) which is probably due to the fact that the subject is right-handed. The movement detector in Figure 6 brings up the movement phase while giving (mainly) negative output to the post movement period. This differentiation is not as decisive as desirable, hence further research has to be pursued to improve on this. Nevertheless a pseudo-online BCI run on the recorded data using a combination of the two classifiers gave the very satisfying result of around 10% error rate. Taking this as a 3 classes choice (left, right, none) this corresponds to an information transmission rate of 29 bits/min. 5 Concluding discussion We gave an outline of our BCI system in the experimental context of voluntary self-paced movements. Our approach has the potential for high bit rates, since (1) it works at a high trial frequency, and (2) classification errors are very low. So far we have used untrained individuals, i.e., improvement can come from appropriate training schemes to shape the brain signals. The two-stage process of first a meta classification whether a movement is about to take place and then a decision between left/right finger movement is very natural and an important new feature of the proposed system. Furthermore, we reject only 0.6% of the trials due to artifacts, so our approach seems ideally suited for the true, highly noisy feedback BCI scenario. Finally, the use of state-of-the-art learning machines enables us not only to achieve high decision accuracies, but also, as a by-product of the classification, the few most prominent features that are found match nicely with physiological intuition: the most salient information can be gained between 230?100 ms before the movement with a focus on C3/C4 area, i.e., over motor cortices, cf. Figure 2. There are clear perspectives for improvement in this BCI approach: our future research activities will therefore focus on (a) projection techniques like ICA, (b) time-series approaches to capture the (non-linear) dynamics of the multivariate EEG signals, and (c) construction of specially adapted kernel functions (SVM or kernel FD) in the spirit of, e.g., [17] to ultimately obtain a BCI feedback system with an even higher bit rate and accuracy. Acknowledgements. We thank S. Harmeling, M. Kawanabe, J. Kohlmorgen, J. Laub, S. Mika, G. R?tsch, R. Vig?rio and A. Ziehe for helpful discussions. References [1] J. J. Vidal, ?Toward direct brain-computer communication?, Annu. Rev. Biophys., 2: 157?180, 1973. [2] N. Birbaumer, N. Ghanayim, T. Hinterberger, I. Iversen, B. Kotchoubey, A. K?bler, J. Perelmouter, E. Taub, and H. Flor, ?A spelling device for the paralysed?, Nature, 398: 297?298, 1999. [3] B. O. Peters, G. Pfurtscheller, and H. Flyvbjerg, ?Automatic Differentiation of Multichannel EEG Signals?, IEEE Trans. Biomed. Eng., 48(1): 111?116, 2001. [4] J. R. Wolpaw, D. J. McFarland, and T. M. Vaughan, ?Brain-Computer Interface Research at the Wadsworth Center?, IEEE Trans. Rehab. Eng., 8(2): 222?226, 2000. [5] W. D. Penny, S. J. Roberts, E. A. Curran, and M. J. Stokes, ?EEG-based cummunication: a pattern recognition approach?, IEEE Trans. Rehab. Eng., 8(2): 214?215, 2000. [6] J. D. Bayliss and D. H. Ballard, ?Recognizing Evoked Potentials in a Virtual Environment?, in: S. A. Solla, T. K. Leen, and K.-R. M?ller, eds., Advances in Neural Information Processing Systems, vol. 12, 3?9, MIT Press, 2000. [7] S. Makeig, S. Enghoff, T.-P. Jung, and T. J. Sejnowski, ?A Natural Basis for Efficient BrainActuated Control?, IEEE Trans. Rehab. Eng., 8(2): 208?211, 2000. [8] W. Lang, O. Zilch, C. Koska, G. Lindinger, and L. Deecke, ?Negative cortical DC shifts preceding and accompanying simple and complex sequential movements?, Exp. Brain Res., 74(1): 99?104, 1989. [9] R. Q. Cui, D. Huter, W. Lang, and L. Deecke, ?Neuroimage of voluntary movement: topography of the Bereitschaftspotential, a 64-channel DC current source density study?, Neuroimage, 9(1): 124?134, 1999. [10] R. Beisteiner, P. Hollinger, G. Lindinger, W. Lang, and A. Berthoz, ?Mental representations of movements. Brain potentials associated with imagination of hand movements?, Electroencephalogr. Clin. Neurophysiol., 96(2): 183?193, 1995. [11] K. P. Bennett and O. L. Mangasarian, ?Robust Linear Programming Discrimination of two Linearly Inseparable Sets?, Optimization Methods and Software, 1: 23?34, 1992. [12] G. R?tsch, T. Onoda, and K.-R. M?ller, ?Soft Margins for AdaBoost?, 42(3): 287?320, 2001. [13] R. O. Duda, P. E. Hart, and D. G. Stork, Pattern Classification, Wiley & Sons, 2nd edn., 2001. [14] S. Mika, G. R?tsch, and K.-R. M?ller, ?A mathematical programming approach to the Kernel Fisher algorithm?, in: T. K. Leen, T. G. Dietterich, and V. Tresp, eds., Advances in Neural Information Processing Systems 13, 591?597, MIT Press, 2001. [15] ?ILOG Solver, ILOG CPLEX 6.5 Reference Manual?,      , 1999. [16] V. Vapnik, The nature of statistical learning theory, Springer Verlag, New York, 1995. [17] K.-R. M?ller, S. Mika, G. R?tsch, K. Tsuda, and B. Sch?lkopf, ?An Introduction to KernelBased Learning Algorithms?, IEEE Transactions on Neural Networks, 12(2): 181?201, 2001.
2030 |@word blankertz1:1 trial:27 neurophysiology:1 briefly:1 norm:2 loading:1 seems:1 duda:1 nd:1 simulation:1 eng:4 pressed:1 cp2:1 contains:1 series:1 chervonenkis:1 tuned:1 franklin:1 past:1 existing:1 reaction:1 current:1 ida:1 lang:3 issuing:1 tetraplegic:1 shape:1 enables:1 motor:20 designed:1 v:4 alone:1 pursued:1 selected:3 device:5 cp3:1 discrimination:1 accordingly:2 xk:5 short:2 filtered:2 mental:1 detecting:1 contribute:1 location:3 c6:1 windowed:1 along:2 c2:1 mathematical:2 differential:1 direct:1 agcl:1 laub:1 behavioral:2 acquired:1 inter:2 ica:1 expected:1 behavior:1 indeed:1 rapid:1 growing:1 multi:2 brain:21 roughly:1 detects:1 automatically:1 actual:2 little:2 str:1 window:7 kohlmorgen:1 becomes:1 ller1:1 solver:1 maximizes:1 panel:1 hemispherical:1 what:3 kind:1 ca1:1 ag:1 differentiation:2 freie:1 pseudo:4 synchronisation:1 every:2 universit:1 classifier:24 scaled:1 demonstrates:1 control:10 makeig:1 before:11 timing:5 referenced:1 vig:1 modulation:1 approximately:1 might:1 chose:3 twice:2 mika:3 evoked:1 relaxing:1 shaded:2 co:1 averaged:1 harmeling:1 spontaneously:1 practice:1 wolpaw:2 procedure:2 danger:1 area:2 empirical:2 crossover:1 significantly:1 thought:1 projection:2 attain:1 pre:1 intention:1 refers:2 downsampled:1 reject:1 get:1 close:2 selection:8 context:3 risk:3 applying:1 vaughan:1 disturbed:1 equivalent:2 map:1 demonstrated:2 center:1 resembling:1 go:1 attention:1 occipital:2 duration:1 unforced:1 his:1 transmit:1 imagine:1 trigger:1 construction:1 user:3 programming:3 edn:1 us:2 curran:1 rfd:3 satisfying:1 recognition:1 located:1 continues:1 electromyogram:1 std:1 predicts:1 capture:1 calculate:1 graz:1 ensures:1 region:1 solla:1 movement:38 trade:2 decrease:1 yk:5 substantial:1 benjamin:2 intuition:1 environment:1 ideally:1 tsch:4 dynamic:1 ultimately:1 trained:5 uniformity:1 technically:1 upon:1 efficiency:1 f2:1 basis:1 neurophysiol:1 various:1 finger:11 regularizer:1 train:1 forced:1 sejnowski:1 detected:2 precedes:1 neurophysics:1 klaus:1 hyper:1 larger:1 bci:17 browser:1 bereitschaftspotential:2 unseen:1 transform:1 exogeneous:1 noisy:2 bler:1 online:11 propose:1 product:1 deecke:2 rehab:3 relevant:2 aligned:3 rapidly:1 achieve:1 pronounced:2 crossvalidation:2 inbuilt:1 electrode:3 transmission:2 produce:1 measured:1 nearest:2 progress:1 strong:1 somatosensory:2 come:1 direction:2 foot:1 thick:2 filter:2 f4:1 subsequently:1 human:2 enable:1 translating:1 virtual:1 bin:2 explains:1 f1:1 generalization:1 suitability:1 elevation:1 strictly:1 accompanying:1 around:3 normal:1 exp:1 mapping:1 optimizer:1 consecutive:1 inseparable:1 conceiving:1 purpose:1 favorable:1 applicable:1 label:3 healthy:1 curio2:1 electroencephalogr:1 sfd:7 mit:2 interfacing:1 aim:1 rather:2 command:7 unspecific:1 derived:1 focus:3 properly:1 improvement:2 indicates:1 mainly:1 contrast:2 sense:1 am:1 posteriori:1 rio:1 helpful:1 nn:3 typically:1 initially:1 relation:2 cp6:1 selects:1 germany:3 biomed:1 issue:2 classification:38 priori:1 art:2 spatial:1 neuen:1 wadsworth:1 field:1 once:1 f3:1 nicely:1 having:2 shaped:1 optimising:1 afc:2 thin:1 mimic:1 future:1 report:2 stimulus:1 contaminated:1 serious:1 others:1 employ:1 distinguishes:1 few:1 bodily:1 individual:2 phase:4 cplex:2 detection:2 interest:1 fd:5 highly:5 investigate:2 intra:1 evaluation:1 inbetween:1 yielding:1 predefined:1 paralysed:2 tuple:1 capable:1 respective:1 euclidean:2 circle:1 tsuda:1 re:1 obscured:1 uncertain:1 tongue:1 increased:1 handed:1 earlier:1 soft:3 kekul:1 ca3:1 recognizing:1 successful:1 conducted:1 optimally:1 stored:1 perelmouter:1 emg:9 jittered:1 accomplish:1 density:1 international:1 grand:1 discriminating:1 ghanayim:1 off:2 picking:1 imagery:3 central:2 recorded:3 tube:2 vigilance:1 perc:1 positivity:1 henceforth:1 hinterberger:1 imagination:3 potential:9 converted:1 de:1 sec:3 coefficient:1 onset:4 decisive:1 later:1 break:2 view:1 performed:2 endogenous:1 doing:1 reached:1 competitive:1 hindenburgdamm:1 start:1 option:1 capability:1 contribution:1 ass:1 minimize:1 square:1 accuracy:7 responded:1 variance:4 yield:2 blink:1 sensori:1 generalize:1 lkopf:1 raw:6 manages:1 therewith:1 none:4 mc:4 kotchoubey:1 detector:2 reach:1 manual:2 ed:2 definition:1 failure:1 sensorimotor:2 frequency:3 involved:1 acquisition:1 associated:2 monitored:1 gain:1 auditory:1 sampled:1 logical:1 knowledge:1 dimensionality:1 positioned:1 actually:2 reflecting:1 back:2 higher:1 day:1 adaboost:1 response:8 erd:1 formulation:2 done:5 though:1 leen:2 furthermore:1 just:1 stage:1 bilaterally:1 until:1 hand:6 horizontal:1 brings:1 artifact:5 reveal:1 bcis:1 laterality:2 effect:1 dietterich:1 concept:2 true:2 consisted:1 regularization:3 hence:2 spatially:1 during:1 self:6 rhythm:1 percentile:3 coincides:1 m:21 prominent:1 palais:1 outline:2 electroencephalogram:1 interface:2 interpreting:1 percent:1 cp4:1 mangasarian:1 common:1 specialized:1 overlearned:1 cpz:1 stork:1 birbaumer:2 endpoint:1 cerebral:1 discussed:2 belong:1 resting:1 measurement:2 significant:1 taub:1 tuning:1 automatic:1 session:2 similarly:1 language:1 stable:1 cortex:8 impressive:1 similarity:1 etc:1 multivariate:2 perspective:1 moderate:1 driven:1 hemisphere:1 scenario:3 keyboard:3 certain:2 initiation:1 verlag:1 meta:1 binary:3 success:1 refrain:1 minimum:1 greater:2 relaxed:1 preceding:1 employed:1 maximize:2 period:2 ller:4 signal:27 sliding:3 full:1 multiple:1 desirable:1 enghoff:1 exceeds:1 instruct:1 match:1 long:1 lin:1 concerning:2 post:1 hart:1 variant:5 involving:1 regression:1 patient:4 metric:1 kernel:4 cz:1 c1:1 addition:1 addressed:1 interval:1 median:4 source:1 crucial:1 sch:1 electrooculogram:1 rest:3 specially:1 posse:1 flor:1 probably:1 subject:18 hz:13 veto:1 spirit:1 split:1 enough:1 easy:1 fft:1 wn:1 gave:2 competing:1 regarding:1 shift:1 favour:2 unprocessed:2 whether:2 unnatural:1 effort:1 peter:1 york:1 speaking:1 cp5:1 repeatedly:1 matlab:1 proceed:1 gabriel:1 useful:1 clear:1 maybe:1 band:6 induces:1 svms:4 processed:2 multichannel:1 fz:1 deliberate:1 cp1:1 shifted:1 sign:1 estimated:1 per:1 vol:1 group:1 key:6 four:1 salient:1 nevertheless:1 drawn:1 preprocessed:2 regulization:1 libsvm:1 button:1 merely:2 run:1 jitter:1 flexor:1 ca2:1 arrive:1 place:1 separation:1 decision:6 comparable:2 bit:7 layer:1 internet:1 bound:1 paced:5 distinguish:1 display:1 correspondence:1 fold:2 yielded:1 activity:4 scalp:5 strength:1 adapted:2 placement:1 precisely:1 constraint:1 comparision:1 bp:6 software:1 speed:2 min:5 chair:1 concluding:1 performing:1 format:1 combination:2 cui:1 smaller:1 son:1 appealing:1 rev:1 berthoz:1 restrains:1 taken:1 eventually:1 mechanism:1 know:1 fed:1 available:3 vidal:1 apply:1 observe:1 kawanabe:1 appropriate:1 alternative:3 denotes:1 cf:2 opportunity:1 log2:3 iversen:1 clin:1 pushing:1 calculating:2 exploit:1 giving:1 graded:4 upcoming:2 seeking:1 objective:2 intend:1 question:1 already:1 primary:2 spelling:1 surrogate:1 klinikum:1 distance:2 thank:1 berlin:3 lateral:1 majority:2 separating:1 whom:1 discriminant:7 reason:2 toward:1 besides:1 o1:1 index:2 length:1 kk:1 ratio:1 downsampling:1 minimizing:3 setup:3 executed:1 robert:2 localised:1 negative:3 trace:2 implementation:3 reliably:3 perform:1 vertical:2 observation:1 ilog:2 displayed:1 voluntary:5 situation:3 extended:1 variability:1 communication:1 stokes:1 dc:3 smoothed:1 arbitrary:1 competence:1 c3:5 c4:3 recalled:1 potsdam:2 learned:1 distinction:2 trans:4 able:1 mcfarland:1 usually:1 pattern:3 challenge:1 program:1 built:1 max:1 power:2 event:11 overlap:1 natural:5 typing:4 regularized:2 rely:1 arm:1 scheme:1 improve:1 brief:2 eye:1 fraunhofer:1 negativity:1 coupled:1 extract:1 eog:1 tresp:1 review:1 prior:3 acknowledgement:1 relative:4 loss:2 expect:1 topography:1 generation:1 interesting:2 filtering:1 mounted:1 degree:2 consistent:1 classifying:2 heavy:1 translation:1 periodicity:1 course:1 elsewhere:1 jung:1 last:2 free:1 keeping:1 offline:1 allow:1 neighbor:2 focussed:1 taking:3 sparse:5 penny:1 feedback:4 dimension:3 cortical:3 withhold:1 calculated:3 ending:1 autoregressive:2 sensory:1 c5:1 adaptive:1 preprocessing:10 projected:4 far:1 cope:1 transaction:1 approximate:1 keep:2 dealing:1 categorised:1 unreliable:1 correlating:1 active:1 overfitting:2 sat:1 investigating:1 consuming:1 discriminative:2 neurology:1 continuous:1 table:4 flyvbjerg:1 ballard:1 nature:2 channel:14 learn:1 transfer:1 robust:2 onoda:1 eeg:20 investigated:2 complex:2 untrained:1 linearly:1 border:1 noise:3 subsample:1 convey:1 bayliss:1 depicts:1 slow:2 wiley:1 pfurtscheller:2 formalization:1 neuroimage:2 position:3 explicit:1 minute:2 keystroke:9 down:1 annu:1 specific:2 er:1 physiological:2 svm:2 vapnik:2 sequential:1 gained:1 magnitude:1 execution:3 biophys:1 cursor:3 margin:5 rejection:2 suited:1 depicted:1 likely:1 visual:1 set2:1 chang:1 springer:1 ch:1 corresponds:1 determines:1 chance:2 sorted:1 marked:1 lvq:1 loosing:1 goal:2 towards:3 fisher:8 considerable:1 bennett:1 determined:3 specifically:2 typical:3 hyperplane:1 total:1 called:1 pas:4 engaged:1 experimental:6 caz:1 preand:1 ziehe:1 support:3 scp:1 latter:1 kernelbased:1 frontal:2 preparation:5 dept:1
1,131
2,031
Correlation Codes in Neuronal Populations Maoz Shamir and Haim Sompolinsky Racah Institute of Physics and Center for Neural Computation, The Hebrew University  of  Jerusalem,    Jerusalem     91904, Israel Abstract Population codes often rely on the tuning of the mean responses to the stimulus parameters. However, this information can be greatly suppressed by long range correlations. Here we study the efficiency of coding information in the second order statistics of the population responses. We show that the Fisher Information of this system grows linearly with the size of the system. We propose a bilinear readout model for extracting information from correlation codes, and evaluate its performance in discrimination and estimation tasks. It is shown that the main source of information in this system is the stimulus dependence of the variances of the single neuron responses. 1 Introduction Experiments in the last years have shown that in many cortical areas, the fluctuations in the responses of neurons to external stimuli are significantly correlated [1, 2, 3, 4], raising important questions regarding the computational implications of neuronal correlations. Recent theoretical studies have addressed the issue of how neuronal correlations affect the efficiency of population coding [4, 5, 6]. It is often assumed that the information about stimuli is coded mainly in the mean neuronal responses, e.g., in the tuning of the mean firing rates, and that by averaging the tuned responses across large populations, an accurate estimate can be obtained despite the significant noise in the single neuron responses. Indeed, for uncorrelated neurons the Fisher Information of the population is extensive [7]; namely, it increases linearly with the number of neurons in the population. Furthermore, it has been shown that this extensive information can be extracted by relatively simple linear readout mechanisms [7, 8]. However, it was recently shown [6] that positive correlations which vary smoothly with space may drastically suppress the information in the mean responses. In particular, the Fisher Information of the system saturates to a finite value as the system size grows. This raises questions about the computational utility of neuronal population codes. Neuronal population responses can represent information in the higher order statistics of the responses [3], not only in their means. In this work, we study the accuracy of coding information in the second order statistics. We call such schemes correlation codes. Specifically, we assume that the neuronal responses obey multivariate Gaussian statistics governed by a stimulus-dependent correlation matrix. We ask whether the Fisher Information of such a system is extensive even in the presence of strong correlations in the neuronal noise. Secondly, we inquire how information in the second order statistics can be efficiently extracted. 2 Fisher Information of a Correlation Code       Our model consists of a system of neurons that code a 2D angle , . Their stochastic response is given by a vector of activities where is the activity of the -th neuron in the presence of a stimulus , and is distributed according to a multivariate Gaussian distribution    "! $#&% )' (   (+*  ,.-0/  21   (3*  04 (1)    Here  is the mean activity of the -th  neuron and its dependence on  is usually referred to as the tuning curve of the neuron; /  is the correlation matrix; and is a normalization constant. Here we shall limit ourselves to the case of multiplicative modulation of the correlations. Specifically we use 5 7 6   8 9  09 6  :<5 ; =6 (2) ( >6 ( ( (  > 5 ; 76  5 ; ?> > 6 @ 76)A    @ =6  !B#%DC (3) E F ( ( K> (4) 9  8 9 ?> : G !$#&% CGHBIJ LNM  F  where and E are the correlation strength and correlation length respectively; L defines the > tuning width the angle at which the variance of the -th  of the correlations; and denotes is shown in Fig. 1. It is important to note that the neuron, 9 M  , is maximal. An 5 example 76 variance adds a contribution to which is larger than the contribution of the smooth part of the correlations. For reasons that will become clear below, we write, 5 =O 6 where 5 7 6  : G 5P 7O 6  : A P5 Q   .@ =6 5 Q denotes the smooth part of the correlation matrix and diagonal part, which in the example of Eqs. (2)-(4) is 5P Q  G   (  ,9 M    (5) the discontinuous (6) A useful measure of the accuracy of a population code is the Fisher Information (FI). In the case of uncorrelated populations it is well known that FI increases linearly with system size [7], indicating that the accuracy of the population coding improves as the system size is increased. Furthermore, it has been shown that relatively simple, linear schemes can provide reasonable readout models for extracting the information in uncorrelated populations [8]. In the case of a correlated multivariate Gaussian distribution, FI is given as , where R  RSUT.V2W A R:XZY0[0[ RSUT.V2W  *  .\ - /  : 1  *  ,\ (7)    ^ R:XZY0[0[     M (8) * * N] /  1 / 0.\_ where \ and / \ denote derivatives of and / with respect to  , respectively. The form *  roles. First they control the of these terms reveals that in general the correlations play two : (note the dependence of efficiency of the information encoded in the mean activities R SUT.V2W on 5 ). Secondly, /   provides an additional source of information  bdstimc e , it R&XZY0[0[ ). When the correlations are independent of the stimulus,  `  9 about  a the ulus ( gf E was shown [6] that positive correlations,  , with long correlation length, ih    , ?=0o C(?,?) ?=?60o o ?=60 ?=?120o ?180 5 K>  ?120 ?=120o ?60 ? 0 >  > (  60 120 180 [deg] (  >  E   Figure 1: The stimulus-dependent correlation matrix, Eqs. (2)-(4), depicted    as a function of two angles, , where and . Here, , and  . L   i cause the saturation of FI to a finite limit at large . This implies that in the presence of such correlations, population averaging cannot overcome the noise even in large networks. This analysis however, [6], did not take into account stimulus-dependent correlations, which is the topic of the present work. RNXZY,[ [ , Eq. (8), we find it useful to write R XZY0[0[  R Q A R O (9) where  R Q    C  5 5 Q Q   , \ M (10)    F 5 Q , ( stimulus-dependent is FI of an uncorrelated population with variance which equals R : R Z X 0 Y 0 [ [ R Q . Evaluating and scales linearly with ; R O  terms for the multiplicative R  R Q these  O . Furthermore, numerical evaluamodel, Eq. (2), we find that R is positive, so that tion of this term shows that O saturates at large to a small finite value, so that for large R:XZY0[ [  R Q   M > C 9 \ ??> >  M (11) F  9  R XZY0[0[ increases linearly with and is equal, for as shown in Fig. 2. We thus conclude that R large , to the FI of variance coding namely to of an independent population in which Analyzing the dependence of information is encoded in their activity variances. Since in our system the information is encoded in the second order statistics of the population responses, it is obvious that linear readouts are inadequate. This raises the question of whether there are relatively simple nonlinear readout models for such systems. In the next sections we will study bilinear readouts and show that they are useful models for extracting information from correlation codes. 3 A Bilinear Readout for Discrimination Tasks    M  A @  Ina two-interval discrimination task the system is given two sets of neuronal activities   generated by two proximal stimuli and and must infer which stimulus generated which activity. The Maximum-Likelihood (ML) discrimination yields the  ?3 1 1 x 10 ?2 [deg ] ?2 [deg ] 0.8 0.6 s 0.5 J Jcorr 0.4 0.2 0 0 200 400 600 800 N 1000 0 0 200 400 600 800 1000 N R XZY0[0[ Figure 2: (a) Fisher Information, , of the stimulus-dependent correlations, Eqs. (2)(4), as a function of the number of neurons in the system. In (b) we show the difference between the  full FI and the contribution of the diagonal term, - as defined by Eq. (9).  Here , and  . Note the different scales in (a) and (b).  E   L  probability of error given by  discriminability equals  \ RO  \     , where       d 1   M   \   @  R     `1   M and the (12)  It has been previously shown that in the case of uncorrelated populations with mean coding, the optimal linear readouts achieves the Maximum-Likelihood discrimination performance in large N [7]. * In order to isolate the properties of correlation coding we will assume that no information is coded in the average firing rates of the neurons, and take  hereafter. We suggest a bilinear readout as a simple generalization of the linear readout to correlation codes. In a discrimination task the bilinear readout makes a decision according to the sign of              (13)  ( A   where a  =6  76  6  ( M 6 M    A @  76   5 1    7 \ 6    decision refers to . Maximizing the signal-to-noise ratio of this rule, the optimal bilinear discriminator (OBD) matrix is given by (14) Using the optimal weights to evaluate the discrimination error we obtain that in large the performance of the OBD saturates the ML performance, Eq. (12). Thus, since FI of this model increases linearly with the size of the system, the discriminability increases as . /  Since the correlation matrix depends on the stimulus, , the OBD matrix, Eq. (14), will also be stimulus dependent. Thus, although the OBD is locally efficient, it cannot be used as such as a global efficient readout. 4 A Bilinear Readout for Angle Estimation 4.1 Optimal bilinear readout for estimation To study the global performance of bilinear readouts we investigate bilinear readouts which minimize the square error of estimating the angle averaged over the whole  range of . For convenience we use complex notation for the encoded angle, and write  as the estimator  of  `   . Let  76   76   =6 K 26 (15) where  are stimulus independent complex weights. We define the optimal bilinear estimator (OBE) as the set of weights that minimizes on average the quadratic estimation error of an unbiased estimator. This error is given by  G     @   M (           is the Lagrange multiplier of the constraint               (16)   where . In general,  it is impossible to find a perfectly unbiased estimator for a continuously varied stimulus, using a finite number of weights. However, in the case of angle estimation, we can employ the underlying rotational symmetry to generate such an estimator. For this we use the  this case one can show that the Lagrange symmetry of the correlation matrix, Eq. (2). In multipliers have the simple form of , and the OBE weight matrix is in the form   of  ) 76  ?> ( ?> ) ( ?> ` > 6  !B#%DC > A > 6 (17) F  (  A  d . This form of a readout matrix, Eq. (17), ?> G  >  and where K>  can guarantees that the estimator will be unbiased. Using these symmetry properties, be written in the following form (for even )   M 1  W  (  > (  ^ b A ?>   @ > " (18) W  HBIJ  _ K>  . These numerical results (Fig. 3 (a)) Figure 3 (a) presents an exampleK>of the function also suggest > that the function  is mainly determined by a few harmonics plus a delta peak at   . Below we will use this fact to study simpler forms of bilinear readout.              Further analysis of the OBE performance in the large totic result           @ : M 1    Q 5 ?> M Q Q  5 ?>  M Q 0 M  limit yields the following asymp- M   c  >  , ( `     (19) Figure 3 (b) shows the numerical calculation of the OBE error (open circles) as a function of . The dashed line is the asymptotic behavior, given by Eq. (19). The dotted line is the Creamer-Rao bound. From the graph one can see that the estimation efficiency of this readout grows linearly with the size of the system, , but is lower than the bound. 4.2 Truncated bilinear readout K>  Motivated by the simple structure of the optimal readout matrix observed in Fig. 3 (a), we studied a bilinear readout of the form of Eqs. (17) and (18) with  which has a delta function peak at the origin plus a few harmonics. Restricting the number of harmonics to  relatively small integers, we evaluated numerically the optimal values of the coefficients   for large systems. Surprisingly we found that for small  and large , these coef ficients approach a value which is independent of the specifics of the model and equals    , yielding a bilinear weight matrix of the form W W  ( B  76   @ =6 (  ^  b (   K> ( > 6  _ !B#%DC > A > 6 W  HBIJ   F     (20) Figure 4 shows the numerical results for the squared average error of this readout for several values of  ! and . The results of Fig. 4 show that for a given  the  8    0.1 (b) (a) w(?) J ???2 0.05 ?2 [deg ] 0 ?2 0 0 0 2 ? 100 200 300 400 N ?>    Figure 3: (a) Profile of  , Eq. (17), for the OBE with . (b) Numerical evaluation of one over the squared estimation error, for the optimal bilinear readout in the multiplicative modulation model (open circles). The dashed line is the asymptotic behavior, given by   Eq. (19). Here , for the optimal bilinear readout in the  multiplicative  modulation model. The dotted line is the FI bound. In these simulations ,  and were used.      @ & M  M L    OV-  E   inverse square error initially increases linearly with but saturates in the limit of large . However, the saturation size increases rapidly with  . The precise form of   depends on the specifics of the correlation model. For the exponentially decaying  . Figure 4 shows that for this range of correlations assumed in Eq. (2), we find , and  the deviations of the inverse square error from linearity are small. Thus, in the regime   ,  is given by the asymptotic behavior, Eq. (19),    shown by the dashed line. OV-   OV- O V -    @ : M We thus conclude that the OBE (with unlimited  ) will generate an inverse square estimawith a coefficient given by Eq. (19), and that tion error which increases linearly with this value can be achieved for reasonable values of by an approximate bilinear weight matrix, of the form of Eq. (20), with small  . The asymptotic result, Eq. (19), is smaller than the optimal value given by the full FI, Eq. (11), see Fig. 4 (dotted line). In fact, it is equal to the error of an independent population with a variance which equals and a quadratic population vector readout of the form     5 Q     M `    (21) It is important to note that in the presence of correlations, the quadratic readout of Eq. (21) is very inefficient, yielding a finite error for large as shown in Fig. 4 (line marked ?quadratic?). 5 Discussion To understand the reason for the simple form of the approximately optimal bilinear weight  matrix, Eq. (20), we rewrite Eq. (15) with of Eq. (20) as   ? `         @ 7 6 (   60 W    (22) ` W   1    1      26 (23) 0.5 1/ ??2 deg?2 0.4 J p=3 0.3 p=2 0.2 p=1 0.1 quadratic 0 0 500 1000 1500 2000 N      E   L  Figure 4: Inverse square estimation error ofthe   finite- approximation for the OBE, Eq. (20). Solid curves from the bottom  . The bottom curve is  . The dashed line is the asymptotic behavior, given by Eq. (19). The FI bound is shown by the dotted    , and were used. line. For the simulations   Comparing this form with Eq. (21) it can be seen that our readout is in the form of a bilinear population vector in which the lowest Fourier modes of the response vector have been removed. Retaining only the high Fourier modes in the response profile suppresses the cross-correlations between the different components of the residual responses  because the underlying correlations have smooth spatial dependence, whose power is concentrated mostly in the low Fourier modes. On the other hand, the information contained in the variance is not removed because the variance contains a discontinuous spatial component, . In other words, the variance of a correlation profile which has only high Fourier modes can still preserve its slowly varying components. Thus, by projecting out the low Fourier modes of the spatial responses the spatial correlations are suppressed but the information in the response variance is retained. 5 Q    This interpretation of the bilinear readout implies that although all the elements of the correlation matrix depend on the stimulus, only the stimulus dependence of the diagonal elements is important. This important conclusion is borne out by our theoretical results concerning the performance of the system. As Eqs. (11) and (19) show, the asymptotic performance of both the full FI as well as that of the OBE are equivalent to those of an . uncorrelated population with a stimulus dependent variance which equals 5 Q    Although we have presented results here concerning a multiplicative model of correlations, we have studied other models of stimulus dependent correlations. These studies indicate that the above conclusions apply to a broad class of populations in which information is encoded in the second order statistics of the responses. Also, for the sake of clarity we have assumed here that the mean responses are untuned,  . Our studies have shown that adding tuned mean inputs does not modify the picture since the smoothly varying positive correlations greatly suppress the information embedded in the first order statistics. *  The relatively simple form of the readout Eq. (22) suggests that neuronal hardware may be able to extract efficiently information embedded in local populations of cells whose noisy responses are strongly correlated, provided that the variances of their responses are significantly tuned to the stimulus. This latter condition is not too restrictive, since tuning of variances of neuronal firing rates to stimulus and motor variables is quite common in the nervous system. Acknowledgments This work was partially supported by grants from the Israel-U.S.A. Binational Science Foundation and the Israeli Science Foundation. M.S. is supported by a scholarship from the Clore Foundation. References [1] E. Fetz, K. Yoyoma and W. Smith, Cerebral Cortex (Plenum Press, New York, 1991). [2] D. Lee, N.L. Port, W. Kruse and A.P. Georgopoulos, J. Neurosci.  , 1161 (1998). [3] E.M. Maynard, N.G. Hatsopoulos, C.L. Ojakangas, B.D. Acuna, J.N. Sanes, R.A. Normann, and J.P. Donoghue, J. Neurosci. 19, 8083 (1999). [4] E. Zohary, M.N. Shadlen and W.T. Newsome, Nature   , 140 (1994). [5] L.F. Abbott and P. Dayan, Neural Computation  , 91 (1999). [6] H. Sompolinsky, H. Yoon, K. Kang and M. Shamir, Phys. Rev. E,  , 051904 (2001); H. Yoon and H. Sompolinsky, Advances in Neural Information Processing Systems 11 (pp. 167). Kearns M.J, Solla S.A and Cohn D.A, Eds., (Cambridge, MA: MIT Press, 1999). [7] S. Seung and H. Sompolinsky, Proc. Natl. Acad. Sci. USA  , 10794 (1993). [8] E. Salinas and L.F. Abbott, J. Comp. Neurosci.  , 89 (1994).
2031 |@word open:2 simulation:2 solid:1 ulus:1 contains:1 hereafter:1 tuned:3 comparing:1 must:1 written:1 numerical:5 motor:1 discrimination:7 nervous:1 smith:1 provides:1 simpler:1 become:1 consists:1 indeed:1 behavior:4 zohary:1 provided:1 estimating:1 notation:1 underlying:2 linearity:1 lowest:1 israel:2 sut:3 minimizes:1 suppresses:1 guarantee:1 ro:1 control:1 grant:1 positive:4 local:1 modify:1 limit:4 acad:1 bilinear:21 despite:1 analyzing:1 fluctuation:1 firing:3 modulation:3 approximately:1 plus:2 discriminability:2 studied:2 suggests:1 range:3 averaged:1 acknowledgment:1 area:1 significantly:2 word:1 refers:1 suggest:2 acuna:1 cannot:2 convenience:1 impossible:1 equivalent:1 center:1 maximizing:1 jerusalem:2 rule:1 estimator:6 racah:1 population:24 plenum:1 shamir:2 play:1 origin:1 element:2 totic:1 observed:1 role:1 bottom:2 p5:1 yoon:2 inquire:1 readout:30 sompolinsky:4 solla:1 removed:2 hatsopoulos:1 seung:1 raise:2 ov:3 rewrite:1 depend:1 efficiency:4 ojakangas:1 salina:1 whose:2 encoded:5 larger:1 quite:1 statistic:8 noisy:1 propose:1 maximal:1 rapidly:1 maoz:1 clore:1 eq:29 strong:1 implies:2 indicate:1 discontinuous:2 stochastic:1 generalization:1 secondly:2 obe:8 vary:1 achieves:1 estimation:8 proc:1 mit:1 gaussian:3 varying:2 likelihood:2 mainly:2 greatly:2 dependent:8 dayan:1 initially:1 issue:1 retaining:1 spatial:4 equal:7 broad:1 stimulus:23 employ:1 few:2 preserve:1 ourselves:1 investigate:1 evaluation:1 yielding:2 natl:1 implication:1 accurate:1 asymp:1 circle:2 theoretical:2 hbi:2 increased:1 rao:1 newsome:1 deviation:1 inadequate:1 too:1 proximal:1 peak:2 lee:1 physic:1 continuously:1 squared:2 slowly:1 borne:1 external:1 derivative:1 inefficient:1 account:1 coding:7 coefficient:2 depends:2 multiplicative:5 tion:2 decaying:1 contribution:3 minimize:1 square:5 accuracy:3 variance:14 efficiently:2 yield:2 comp:1 phys:1 coef:1 ed:1 pp:1 obvious:1 ask:1 improves:1 higher:1 response:22 evaluated:1 strongly:1 furthermore:3 correlation:41 hand:1 cohn:1 nonlinear:1 maynard:1 defines:1 v2w:3 mode:5 grows:3 usa:1 unbiased:3 multiplier:2 width:1 harmonic:3 recently:1 fi:12 common:1 binational:1 exponentially:1 cerebral:1 interpretation:1 numerically:1 significant:1 cambridge:1 tuning:5 cortex:1 add:1 multivariate:3 recent:1 seen:1 additional:1 kruse:1 signal:1 dashed:4 full:3 infer:1 smooth:3 calculation:1 cross:1 long:2 concerning:2 coded:2 represent:1 normalization:1 achieved:1 cell:1 addressed:1 interval:1 source:2 isolate:1 call:1 extracting:3 integer:1 presence:4 affect:1 perfectly:1 regarding:1 donoghue:1 whether:2 motivated:1 utility:1 york:1 cause:1 useful:3 clear:1 locally:1 concentrated:1 hardware:1 generate:2 dotted:4 sign:1 delta:2 lnm:1 write:3 shall:1 clarity:1 abbott:2 graph:1 year:1 angle:7 inverse:4 reasonable:2 decision:2 bound:4 haim:1 quadratic:5 activity:7 strength:1 constraint:1 georgopoulos:1 unlimited:1 sake:1 fourier:5 relatively:5 according:2 across:1 smaller:1 suppressed:2 rev:1 projecting:1 previously:1 mechanism:1 apply:1 obey:1 denotes:2 restrictive:1 scholarship:1 question:3 dependence:6 diagonal:3 sci:1 topic:1 reason:2 code:10 length:2 retained:1 ratio:1 rotational:1 hebrew:1 mostly:1 suppress:2 neuron:12 finite:6 truncated:1 saturates:4 precise:1 dc:3 varied:1 namely:2 extensive:3 discriminator:1 raising:1 kang:1 israeli:1 able:1 usually:1 below:2 regime:1 saturation:2 power:1 rely:1 residual:1 scheme:2 picture:1 extract:1 normann:1 gf:1 asymptotic:6 embedded:2 untuned:1 foundation:3 shadlen:1 port:1 uncorrelated:6 obd:4 surprisingly:1 last:1 supported:2 drastically:1 understand:1 institute:1 fetz:1 distributed:1 overcome:1 curve:3 cortical:1 evaluating:1 approximate:1 deg:5 ml:2 global:2 reveals:1 assumed:3 conclude:2 nature:1 symmetry:3 complex:2 did:1 main:1 linearly:9 neurosci:3 whole:1 noise:4 profile:3 neuronal:11 fig:7 referred:1 sanes:1 governed:1 specific:2 ih:1 restricting:1 adding:1 smoothly:2 depicted:1 lagrange:2 contained:1 partially:1 extracted:2 ma:1 marked:1 fisher:7 specifically:2 determined:1 averaging:2 kearns:1 indicating:1 ficients:1 latter:1 evaluate:2 correlated:3
1,132
2,032
Stochastic Mixed-Signal VLSI Architecture for High-Dimensional Kernel Machines Roman Genov and Gert Cauwenberghs Department of Electrical and Computer Engineering Johns Hopkins University, Baltimore, MD 21218 roman,gert @jhu.edu  Abstract A mixed-signal paradigm is presented for high-resolution parallel innerproduct computation in very high dimensions, suitable for efficient implementation of kernels in image processing. At the core of the externally digital architecture is a high-density, low-power analog array performing binary-binary partial matrix-vector multiplication. Full digital resolution is maintained even with low-resolution analog-to-digital conversion, owing to random statistics in the analog summation of binary products. A random modulation scheme produces near-Bernoulli statistics even for highly correlated inputs. The approach is validated with real image data, and with experimental results from a CID/DRAM analog array prototype in 0.5 m CMOS.  1 Introduction Analog computational arrays [1, 2, 3, 4] for neural information processing offer very large integration density and throughput as needed for real-time tasks in computer vision and pattern recognition [5]. Despite the success of adaptive algorithms and architectures in reducing the effect of analog component mismatch and noise on system performance [6, 7], the precision and repeatability of analog VLSI computation under process and environmental variations is inadequate for some applications. Digital implementation [10] offers absolute precision limited only by wordlength, but at the cost of significantly larger silicon area and power dissipation compared with dedicated, fine-grain parallel analog implementation, e.g., [2, 4]. The purpose of this paper is twofold: to present an internally analog, externally digital architecture for dedicated VLSI kernel-based array processing that outperforms purely digital approaches with a factor 100-10,000 in throughput, density and energy efficiency; and to provide a scheme for digital resolution enhancement that exploits Bernoulli random statistics of binary vectors. Largest gains in system precision are obtained for high input dimensions. The framework allows to operate at full digital resolution with relatively imprecise analog hardware, and with minimal cost in implementation complexity to randomize the input data. The computational core of inner-product based kernel operations in image processing and pattern recognition is that of vector-matrix multiplication (VMM) in high dimensions:       (1)  with  -dimensional input vector  ,  -dimensional output vector , and    matrix  to elements . In artificial neural networks, the matrix elements correspond   synapses, between neurons. The elements also represent templates  weights, or  in a vector quantizer [8], or support vectors in a support vector machine [9]. In  what follows we concentrate on VMM computation which dominates inner-product based 1 kernel computations for high vector dimensions. 2 The Kerneltron: A Massively Parallel VLSI Computational Array 2.1 Internally Analog, Externally Digital Computation The approach combines the computational efficiency of analog array processing with the precision of digital processing and the convenience of a programmable and reconfigurable digital interface. The digital representation is embedded in the analog array architecture, with inputs presented in bit-serial fashion, and matrix elements stored locally in bit-parallel form:  $#            "  !  #     &  ('! & % &  (2) (3)   # *       & .  ! / & )  , -  % +  &   decomposing (1) into: (4) with binary-binary VMM partials:  # 0# #   !&  ! / &   ! ' 2 1  (5) The key is to compute and accumulate the binary-binary partial products (5) using an analog VMM array, and to combine the quantized results in the digital domain according to (4). # Digital-to-analog conversion at the input interface is inherent in the bit-serial implementation, and row-parallel analog-to-digital converters (ADCs) are used at the output interface ! &  / to quantize . A 512  128 array prototype using CID/DRAM cells is shown in Figure 1 (a). 2.2 CID/DRAM Cell and Array 0# $# # element The unit cell in the analog array combines a CID computational   [12, ! 13] with a DRAM storage element. The cell stores one bit of a matrix element , performs   ! ' !& a one-quadrant binary-binary multiplication of and in (5), and accumulates 1 Radial basis kernels with 354 -norm can also be formulated in inner product format. RS(i)m Vout(i)m M1 M2 M3 CID DRAM (i) w mn x(j) n RS(i)m x(j)n Vout(i)m 0 Vdd/2 Vdd Write 0 Vdd/2 Vdd Compute (a) 0 Vdd/2 Vdd (b)  Figure 1: (a) Micrograph of the Kerneltron prototype, containing an array of       CID/DRAM cells, and a row-parallel bank of  flash ADCs. Die size is    in 0.5 m CMOS technology. (b) CID computational cell with integrated DRAM storage. Circuit diagram, and charge transfer diagram for active write and compute operations.   # $# and indices.# The circuit diagram and operation the result across cells with common of the cell are given in Figure 1 (b). An array of cells thus performs (unsigned) binary   ! ' !&  ! /& multiplication (5) of matrix and vector yielding , for values of in parallel across the array, and values of in sequence over time. The cell contains three MOS transistors connected$# in series as depicted in Figure 1 (b). $# Transistors M1 and M2 comprise a dynamic random-access memory (DRAM) cell, with  !   switch M1 controlled by Row Select signal . When activated, the binary quantity   ! is written in the form of charge (either   or 0) stored under the gate of M2. Transistors M2 and M3 in turn comprise a charge injection device (CID), which by virtue of charge conservation moves electric charge between two potential wells in a non-destructive manner [12, 13, 14]. The charge left under the gate of M2 can only be redistributed between the two CID tran$# is sistors, M2 and M3. An active charge transfer from M2 to M3 can only occur if there # non-zero charge stored, and if the potential on the gate of M2 drops below that of M3! [12].  This and ' ! & condition implies a logical AND, i.e., unsigned binary multiplication, of . The multiply-and-accumulate operation is then completed by capacitively sensing the amount of charge transferred onto the electrode of M3, the output summing node. To this end, the voltage on the output line, left floating after being pre-charged to  ,  is observed. When the charge transfer is active, the cell contributes a change in voltage      "! where# #"! is the total capacitance on the output line across cells. The total response is thus ' ! & proportional to the number of actively transferring cells. After deactivating the input , the transferred charge returns to the storage node M2. The CID computation is non-destructive and intrinsically reversible [12], and DRAM refresh is only required to counteract junction and subthreshold leakage. The bottom diagram in Figure 1 (b) depicts the charge transfer timing diagram for write # 0# and compute operations in the case when both   ! ' !& and are of logic level 1. 2.3 System-Level Performance Measurements on the 512  128-element analog array and other fabricated prototypes show a dynamic range of 43 dB, and a computational cycle of 10 s with power consumption of  1 . 50 nW per cell. The size of the CID/DRAM cell is 8  45 with    The overall system resolution is limited by the precision in the quantization of the outputs from the analog array. Through digital postprocessing, two bits are gained over the resolution of the ADCs used [15], for a total system resolution of 8 bits. Larger resolutions can be obtained by accounting for the statistics of binary terms in the addition, the subject of the next section. 3 Resolution Enhancement Through Stochastic Encoding # be achieved (as if computed Since the analog inner product (5) is discrete, zero error can digitally) by matching the quantization levels of the ADC with each of the   discrete  ! / & levels in the inner product. Perfect reconstruction of from the quantized output, for .   bits, assumes the combined effect of noise and an overall resolution of nonlinearity in the analog array and the ADC is within one LSB (least significant bit). For large arrays, this places stringent requirements on analog precision and ADC resolution, .   .             The implicit assumption is that all quantization levels are (equally) needed. A straightforward study of the statistics of the inner product, below, reveals that this is poor use of available resources. 3.1 Bernoulli Statistics #  0#  In what follows   signed, ' ! &  we assume !  rather than unsigned, binary values for inputs and weights,  and  . This translates to exclusive-OR (XOR), rather than AND, multiplication on the analog array, an operation that can be easily accomplished with the CID/DRAM # architecture by differentially coding input and stored bits using twice the number $of columns and unit cells.  # # # $# ' !& the (XOR) product For input   bits ! ' ! & that are Bernoulli distributed (i.e., fair coin flips),  !  ! /& terms in (5) are Bernoulli distributed, regardless of . Their sum  # thus follows a binomial distribution     ! / &  !  #"  %$'&)(   &   ( (6)   #+* 1$101 *  , which in the Central Limit -,/. approaches a normal # 1 , with & distribution with zero mean and variance  . In other words, for random inputs 10 . in high   dimensions the active range (or standard deviation) of the inner-product is , a factor 10 .  smaller than the full range  . In principle, this allows to relax the effective resolution of the ADC. However, any reduction in conversion range will result in a small but non-zero probability of overflow. In practice, the risk of overflow can be reduced to negligible levels with a few additional bits in the ADC conversion range. An alternative strategy is to use a variable resolution ADC which expands the conversion range on rare occurences of overflow. 2 2 Or, with stochastic input encoding, overflow detection could initiate a different random draw. Inner Product 20 10 0 ?10 ?20 0.2 0.4 0.6 0.8 0.4 0.6 0.8 Output Voltage (V) 50 40 Count 30 20 10 0 (a) 0.2 Output Voltage (V) (b) Figure 2: Experimental results from CID/DRAM analog array. (a) Output voltage on the sense line computing exclusive-or inner product of 64-dimensional stored and presented binary vectors. A variable number of active bits is summed at different locations in the array by shifting the presented bits. (b) Top: Measured output and actual inner product for 1,024 samples of Bernoulli distributed pairs of stored and presented vectors. Bottom: Histogram of measured array outputs. 3.2 Experimental Results While the reduced range of the analog inner product supports lower ADC resolution in terms of number of quantization levels, it requires low levels of mismatch and noise so that the discrete levels can be individually resolved, near the center of the distribution. To verify this, we conducted the following experiment. Figure 2 shows the measured outputs on one row of 128 CID/DRAM cells, configured differentially to compute signed binary (exclusive-OR) inner products of stored and presented binary vectors in 64 dimensions. The scope trace in Figure 2 (a) is obtained by storing all  bits, and shifting a sequence of input bits that differ with the stored bits by  bits. The left and right segment of the scope trace correspond to different selections of active bit locations along the array that are maximally disjoint, to indicate a worst-case mismatch scenario. The measured and actual inner products in Figure 2 (b) are obtained by storing and presenting 1,024 pairs of random binary vectors. The histogram shows a clearly resolved, discrete binomial distribution for the observed analog voltage.   For very large arrays, mismatch and noise may pose a problem in the present implementation with floating sense line. A sense amplifier with virtual ground on the sense line and . feedback capacitor optimized to the  range would provide a simple solution. 10 3.3 Real Image Data Although most randomly selected patterns do not correlate with any chosen template, patterns from the real world tend to correlate, and certainly those that are of interest to kernel computation 3 . The key is stochastic encoding of the inputs, as to randomize the bits presented to the analog array. 3 This observation, and the binomial distribution for sums of random bits (6), forms the basis for the associative recall in a Kanerva memory. 500 500 400 400 Count 600 Count 600 300 300 200 200 100 100 0 ?1000 ?500 0 Inner Product 500 0 ?1000 1000 10 10 9 9 8 8 7 7 6 6 Count 11 Count 11 5 4 3 3 2 2 0 ?1000 0 500 1000 0 500 1000 Inner Product 5 4 1 ?500 1 ?500 0 Inner Product 500 1000 0 ?1000 ?500  #Inner Product ! /& Figure 3: Histograms of partial binary inner products for 256 pairs of randomly selected 32  32 pixel segments of Lena. Left: with unmodulated 8-bit image data for both vectors. Right: with 12-bit modulated stochastic encoding of one of  the two vectors. Top: all bit planes and . Bottom: most significant bit (MSB) plane, .  Randomizing an informative input while retaining the information is a futile goal, and we are content with a solution that approaches the ideal performance within observable bounds, and with reasonable cost in implementation. Given that ?ideal? randomized inputs relax the ADC resolution by  .   bits, they necessarily reduce the wordlenght of the output by the same. To account for the lost bits in the range of the output, it is necessary to increase the range of the ?ideal? randomized input by the same number of bits. 10  . One possible stochastic encoding scheme that restores the range is  -fold oversampling of the input through (digital) delta-sigma modulation. This is a workable solution; however we propose one that is simpler and less costly to implement. For each -bit input compo .  in the range   nent  , pick a random integer , and subtract it to produce    .   a modulated input with    additional bits. It can be shown that for   inner product for  is off at most by   worst-case deterministic inputs the mean of the  . from the origin. The desired inner products for  are retrieved by digitally adding    and  . The random offset  can be chosen once, so the inner products obtained for $# its inner product with the templates can be pre-computed upon initializing or programming the array. The implementation cost is thus limited to component-wise subtraction of  ! and  , achieved using one full adder cell, one bit register, and ROM storage of the  bits for every column of the array.   10    10 Figure 3 provides a proof of principle, using image data selected at random from Lena. 12-bit stochastic encoding of the 8-bit image, by subtracting a random variable in a range 15 times larger than the image, produces the desired binomial distribution for the partial bit inner products, even for the most significant bit (MSB) which is most highly correlated. 4 Conclusions We presented an externally digital, internally analog VLSI array architecture suitable for real-time kernel-based neural computation and machine learning in very large dimensions, such as image recognition. Fine-grain massive parallelism and distributed  . memory, in an array of 3-transistor CID/DRAM cells, provides a throughput of   binary MACS (mul tiply accumulates per second) per Watt of power in a 0.5 m process. A simple stochastic encoding scheme relaxes precision requirements in the analog implementation by one bit for each four-fold increase in vector dimension, while retaining full digital overall system resolution.   Acknowledgments This research was supported by ONR N00014-99-1-0612, ONR/DARPA N00014-00-C0315, and NSF MIP-9702346. Chips were fabricated through the MOSIS service. References [1] A. Kramer, ?Array-based analog computation,? IEEE Micro, vol. 16 (5), pp. 40-49, 1996. [2] G. Han, E. Sanchez-Sinencio, ?A general purpose neuro-image processor architecture,? Proc. of IEEE Int. Symp. on Circuits and Systems (ISCAS?96), vol. 3, pp 495-498, 1996 [3] F. Kub, K. Moon, I. Mack, F. Long, ? Programmable analog vector-matrix multipliers,? IEEE Journal of Solid-State Circuits, vol. 25 (1), pp. 207-214, 1990. [4] G. Cauwenberghs and V. Pedroni, ?A Charge-Based CMOS Parallel Analog Vector Quantizer,? Adv. Neural Information Processing Systems (NIPS*94), Cambridge, MA: MIT Press, vol. 7, pp. 779-786, 1995. [5] Papageorgiou, C.P, Oren, M. and Poggio, T., ?A General Framework for Object Detection,? in Proceedings of International Conference on Computer Vision, 1998. [6] G. Cauwenberghs and M.A. Bayoumi, Eds., Learning on Silicon: Adaptive VLSI Neural Systems, Norwell MA: Kluwer Academic, 1999. [7] A. Murray and P.J. Edwards, ?Synaptic Noise During MLP Training Enhances Fault-Tolerance, Generalization and Learning Trajectory,? in Advances in Neural Information Processing Systems, San Mateo, CA: Morgan Kaufman, vol. 5, pp 491-498, 1993. [8] A. Gersho and R.M. Gray, Vector Quantization and Signal Compression, Norwell, MA: Kluwer, 1992. [9] V. Vapnik, The Nature of Statistical Learning Theory, 2nd ed., Springer-Verlag, 1999. [10] J. Wawrzynek, et al., ?SPERT-II: A Vector Microprocessor System and its Application to Large Problems in Backpropagation Training,? in Advances in Neural Information Processing Systems, Cambridge, MA: MIT Press, vol. 8, pp 619-625, 1996. [11] A. Chiang, ?A programmable CCD signal processor,? IEEE Journal of Solid-State Circuits, vol. 25 (6), pp. 1510-1517, 1990. [12] C. Neugebauer and A. Yariv, ?A Parallel Analog CCD/CMOS Neural Network IC,? Proc. IEEE Int. Joint Conference on Neural Networks (IJCNN?91), Seattle, WA, vol. 1, pp 447-451, 1991. [13] V. Pedroni, A. Agranat, C. Neugebauer, A. Yariv, ?Pattern matching and parallel processing with CCD technology,? Proc. IEEE Int. Joint Conference on Neural Networks (IJCNN?92), vol. 3, pp 620-623, 1992. [14] M. Howes, D. Morgan, Eds., Charge-Coupled Devices and Systems, John Wiley & Sons, 1979. [15] R. Genov, G. Cauwenberghs ?Charge-Mode Parallel Architecture for Matrix-Vector Multiplication,? IEEE T. Circuits and Systems II, vol. 48 (10), 2001.
2032 |@word compression:1 norm:1 nd:1 r:2 accounting:1 pick:1 solid:2 reduction:1 contains:1 series:1 outperforms:1 written:1 refresh:1 john:2 grain:2 informative:1 drop:1 msb:2 selected:3 device:2 plane:2 nent:1 core:2 compo:1 chiang:1 quantizer:2 quantized:2 node:2 location:2 provides:2 deactivating:1 simpler:1 along:1 combine:3 symp:1 manner:1 lena:2 actual:2 circuit:6 what:2 kaufman:1 adc:8 fabricated:2 every:1 expands:1 charge:15 unit:2 internally:3 unmodulated:1 negligible:1 engineering:1 timing:1 service:1 limit:1 despite:1 accumulates:2 encoding:7 modulation:2 signed:2 twice:1 mateo:1 limited:3 range:13 acknowledgment:1 yariv:2 practice:1 lost:1 implement:1 agranat:1 backpropagation:1 area:1 jhu:1 significantly:1 matching:2 imprecise:1 pre:2 radial:1 quadrant:1 word:1 convenience:1 onto:1 selection:1 storage:4 unsigned:3 risk:1 deterministic:1 charged:1 center:1 straightforward:1 regardless:1 resolution:17 occurences:1 m2:9 array:30 gert:2 variation:1 massive:1 programming:1 origin:1 element:8 recognition:3 observed:2 bottom:3 electrical:1 initializing:1 worst:2 connected:1 cycle:1 adv:1 digitally:2 complexity:1 dynamic:2 vdd:6 segment:2 purely:1 upon:1 efficiency:2 basis:2 easily:1 resolved:2 darpa:1 chip:1 joint:2 effective:1 artificial:1 larger:3 relax:2 statistic:6 associative:1 sequence:2 transistor:4 reconstruction:1 redistributed:1 tran:1 product:26 propose:1 subtracting:1 differentially:2 seattle:1 enhancement:2 electrode:1 requirement:2 produce:3 cmos:4 perfect:1 object:1 pose:1 measured:4 edward:1 implies:1 indicate:1 differ:1 concentrate:1 owing:1 stochastic:8 stringent:1 virtual:1 generalization:1 summation:1 ground:1 normal:1 ic:1 nw:1 scope:2 mo:1 purpose:2 proc:3 individually:1 largest:1 mit:2 clearly:1 rather:2 voltage:6 validated:1 neugebauer:2 bernoulli:6 sense:4 integrated:1 transferring:1 vlsi:6 pixel:1 overall:3 retaining:2 restores:1 integration:1 summed:1 comprise:2 once:1 throughput:3 roman:2 vmm:4 inherent:1 few:1 randomly:2 micro:1 floating:2 iscas:1 amplifier:1 detection:2 interest:1 mlp:1 highly:2 multiply:1 workable:1 certainly:1 yielding:1 activated:1 norwell:2 partial:5 necessary:1 poggio:1 wordlength:1 capacitively:1 desired:2 mip:1 minimal:1 column:2 cost:4 mac:1 deviation:1 rare:1 inadequate:1 conducted:1 stored:8 randomizing:1 combined:1 density:3 international:1 randomized:2 off:1 hopkins:1 central:1 containing:1 return:1 actively:1 account:1 potential:2 coding:1 int:3 configured:1 register:1 cauwenberghs:4 parallel:11 xor:2 variance:1 moon:1 correspond:2 subthreshold:1 repeatability:1 vout:2 trajectory:1 processor:2 synapsis:1 ed:3 synaptic:1 energy:1 pp:9 destructive:2 proof:1 gain:1 intrinsically:1 logical:1 recall:1 response:1 maximally:1 implicit:1 adder:1 reversible:1 mode:1 gray:1 effect:2 verify:1 multiplier:1 during:1 maintained:1 die:1 presenting:1 performs:2 dissipation:1 dedicated:2 interface:3 postprocessing:1 image:10 wise:1 common:1 analog:33 m1:3 kluwer:2 accumulate:2 silicon:2 measurement:1 significant:3 cambridge:2 nonlinearity:1 access:1 han:1 retrieved:1 massively:1 store:1 scenario:1 n00014:2 verlag:1 binary:21 success:1 onr:2 fault:1 accomplished:1 morgan:2 additional:2 kub:1 subtraction:1 paradigm:1 signal:5 ii:2 full:5 academic:1 offer:2 long:1 serial:2 equally:1 controlled:1 neuro:1 vision:2 histogram:3 represent:1 kernel:8 oren:1 achieved:2 cell:20 addition:1 fine:2 baltimore:1 diagram:5 lsb:1 howe:1 spert:1 operate:1 subject:1 tend:1 db:1 sanchez:1 capacitor:1 integer:1 near:2 ideal:3 relaxes:1 switch:1 architecture:9 converter:1 inner:23 reduce:1 prototype:4 translates:1 programmable:3 amount:1 locally:1 hardware:1 reduced:2 nsf:1 oversampling:1 delta:1 disjoint:1 per:3 write:3 discrete:4 vol:10 key:2 four:1 micrograph:1 mosis:1 sum:2 counteract:1 place:1 reasonable:1 draw:1 bit:36 bound:1 fold:2 occur:1 ijcnn:2 performing:1 injection:1 relatively:1 format:1 transferred:2 department:1 according:1 watt:1 poor:1 across:3 smaller:1 son:1 wawrzynek:1 mack:1 kanerva:1 resource:1 turn:1 count:5 needed:2 initiate:1 flip:1 gersho:1 end:1 junction:1 operation:6 decomposing:1 available:1 alternative:1 coin:1 gate:3 assumes:1 binomial:4 top:2 completed:1 ccd:3 exploit:1 murray:1 overflow:4 leakage:1 move:1 capacitance:1 quantity:1 randomize:2 strategy:1 exclusive:3 md:1 costly:1 cid:15 enhances:1 consumption:1 rom:1 index:1 trace:2 sigma:1 dram:14 implementation:9 conversion:5 neuron:1 observation:1 pair:3 required:1 optimized:1 nip:1 below:2 pattern:5 mismatch:4 parallelism:1 memory:3 shifting:2 power:4 suitable:2 innerproduct:1 mn:1 scheme:4 technology:2 coupled:1 genov:2 bayoumi:1 multiplication:7 embedded:1 mixed:2 proportional:1 pedroni:2 digital:19 principle:2 bank:1 storing:2 row:4 supported:1 template:3 absolute:1 distributed:4 tolerance:1 feedback:1 dimension:8 world:1 adaptive:2 san:1 correlate:2 observable:1 logic:1 active:6 reveals:1 summing:1 conservation:1 nature:1 transfer:4 ca:1 contributes:1 quantize:1 futile:1 necessarily:1 papageorgiou:1 electric:1 domain:1 microprocessor:1 noise:5 fair:1 mul:1 depicts:1 fashion:1 wiley:1 precision:7 externally:4 reconfigurable:1 sensing:1 offset:1 virtue:1 dominates:1 quantization:5 vapnik:1 adding:1 gained:1 subtract:1 depicted:1 springer:1 environmental:1 ma:4 goal:1 formulated:1 kramer:1 flash:1 twofold:1 content:1 change:1 reducing:1 total:3 experimental:3 m3:6 select:1 support:3 modulated:2 correlated:2
1,133
2,033
Grouping and dimensionality reduction by locally linear embedding Marzia Polito Division of Physics, Mathematics and Astronomy California Institute of Technology Pasadena, CA, 91125 polito @caltech.edu Pietro Perona Division of Engeneering and Applied Mathematics California Institute of Technology Pasadena, CA, 91125 perona@caltech.edu Abstract Locally Linear Embedding (LLE) is an elegant nonlinear dimensionality-reduction technique recently introduced by Roweis and Saul [2]. It fails when the data is divided into separate groups. We study a variant of LLE that can simultaneously group the data and calculate local embedding of each group. An estimate for the upper bound on the intrinsic dimension of the data set is obtained automatically. 1 Introduction Consider a collection of N data points Xi E ]RD. Suppose that , while the dimension D is large, we have independent information suggesting that the data are distributed on a manifold of dimension d < < D. In many circumstances it is beneficial to calculate the coordinates Yi E ]Rd of the data on the lower-dimensional manifold, both because the shape of the manifold may yield some insight in the process that produced the data, and because it is cheaper to store and manipulate the data when it is embedded in fewer dimensions. How can we compute such coordinates? Principal component analysis (PCA) is a classical technique which works well when the data lie close to a flat manifold [1]. Elegant methods for dealing with data that is distributed on curved manifolds have been recently proposed [3, 2]. We study one of them, Locally Linear Embedding (LLE) [2], by Roweis and Saul. While LLE is not designed to handle data that are disconnected, i.e. separated into groups, we show that a simple variation of the method will handle this situation correctly. Furthermore, both the number of groups and the upper bound on the intrinsic dimension of the data may be estimated automatically, rather than being given a-priori. 2 Locally linear embedding The key insight inspiring LLE is that, while the data may not lie close to a globally linear manifold, it may be approximately locally linear, and in this case each point may be approximated as a linear combination of its nearest neighbors. The coefficients of this linear combination carries the vital information for constructing a lower-dimensional linear embedding. More explicitly: consider a data set {Xd i=l...,N E ]RD. The local linear structure can be easily encoded in a sparse N by N matrix W, proceeding as follows. The first step is to choose a criterion to determine the neighbors of each point. Roweis and Saul chose an integer number K and pick, for every point, the K points nearest to it. For each point Xi then, they determine the linear combination of its neighbors which best approximates the point itself. The coefficients of such linear combinations are computed by minimizing the quadratic cost function: f(W) = L N IXi - L WijXj 12 (1) j=1 while enforcing the constraints W ij = 0 if Xj is not a neighbor of Xi , and L:.f=1 Wij = 1 for every i; these constraints ensure that the approximation of Xi ~ Xi = L:.f=1 WijXj lies in the affine subspace generated by the K nearest neighbors of Xi, and that the solution W is translation-invariant . This least square problem may be solved in closed form [2]. The next step consists of calculating a set {Yih=1, ... ,N of points in ]Rd, reproducing as faithfully as possible the local linear structure encoded in W. This is done minimizing a cost function N <I>(Y) = N L IYi - L Wij Yjl2 (2) i=1 j =1 To ensure the uniqueness of the solution two constraint are imposed: translation invariance by placing the center of gravity of the data in the origin, i.e. L:i Yi = 0, and normalized unit covariance of the Yi's, i.e. tt L:~1 Yi Q9 Yi = I. Roweis and Saul prove that <I>(Y) = tr(yT MY), where M is defined as M = (I - wf (I - W). The minimum of the function <I>(Y) for the d-th dimensional representation is then obtained with the following recipe. Given d, consider the d + 1 eigenvectors associated to the d + 1 smallest eigenvalues of the matrix M. Then discard the very first one. The rows of the matrix Y whose columns are given by such d eigenvectors give the desired solution. The first eigenvector is discarded because it is a vector composed of all ones, with 0 as eigenvalue. As we shall see, this is true when the data set is 'connected' . 2.1 Disjoint components In LLE every data point has a set of K neighbors. This allows us to partition of the whole data set X into K -connected components, corresponding to the intuitive visual notion of different 'groups' in the data set. We say that a partition X = UiUi is finer than a partition X = Uj 10 if every Ui is contained in some 10. The partition in K -connected components is the finest : ?............................ 10 20 30 40 50 60 70 60 90 100 -020"'---------:::---;:;;-----O;---;;;------;';c------:::--~ Figure 1: (Top-left) 2D data Xi distributed along a curve (the index i increases from left to right for convenience). (Top-right) Coordinates Yi of the same points calculated by LLE with K = 10 and d = 1. The x axis represents the index i and the y axis represents Yi. This is a good parametrization which recognizes the intrinsically I-dimensional structure of the data. (Bottom-left) As above, the data is now disconnected, i.e. points in different groups do not share neighbors. (Bottom-right) One-dimensional LLE calculated on the data (different symbols used for points belonging to the different groups). Notice that the Yi's are not a good representation of the data any longer since they are constant within each group. partition of the data set such that if two points have at least one neighbor in common, or one is a neighbor of the other, then they belong to the same component. Note that for any two points in the same component, we can find an ordered sequence of points having them as endpoints, such that two consecutive points have at least one neighbor in common. A set is K -connected if it contains only one K-connected component . Consider data that is not K -connected, then LLE does not compute a good parametrization, as illustrated in Figure 1. 2.2 Choice of d. How is d chosen? The LLE method [2] is based on the assumption that d is known. What if we do not know it in advance? If we overestimate d it then LLE behaves pathologically. Let us consider a straight line, drawn in 1R3 . Figure 2 shows what happens if d is chosen equal to 1 and to 2. When the choice is 2 (right) then LLE 'makes up' information and generates a somewhat arbitrary 2D curve. As an effect of the covariance constraint, the representation curves the line, the Figure 2: Coordinates Yi calculated for data Xi distributed along a straight line in ]RD = ]R3 when the dimension d is chosen as d = 1 (Left), and d = 2 (Right). The index i is indicated along the x axis (Left) and along the 2D curve (Right). curvature can be very high, and even locally we possibly completely lose the linear structure. The problem is, we chosed the wrong target dimension. The onedimensional LLE works in fact perfectly (see Figure 2, left). PCA provides a principled way of estimating the intrinsic dimensionality of the data: it corresponds to the number of large singular values of the covariance matrix of the data. Is such an estimate possible with LLE as well? 3 Dimensionality detection: the size of the eigenvalues In the example of Figure 2 the two dimensional representation of the data (d = 2) is clearly the 'wrong' one, since the data lie in a one-dimensional linear subspace. In this case the unit covariance constraint in minimizing the function <I>(Y) is not compatible with the linear structure. How could one have obtained the correct estimate of d? The answer is that d + 1 should be less or equal to the number of eigenvalues of M that are close to zero. Proposition 1. Assume that the data Xi E ]RD is K -connected and that it is locally fiat, i.e. there exists a corresponding set Yi E ]Rd for some d > 0 such that Yi = L: j Wij}j (zero-error approximation), the set {Yi} has rank d, and has the origin as center of gravity: L:~1 Yi = the matrix M. Then d < z . o. Call z the number of zero eigenvalues of Proof. By construction the N vector composed of all 1 's is a zero-eigenvector of M. Moreover, since the Yi are such that the addends of <I> have zero error, then the matrix Y , which by hypothesis has rank d, is in the kernel of I - W and hence in the kernel of M. Due to the center of gravity constraint, all the columns of Y are orthogonal to the all 1 's vector. Hence M has at least d + 1 zero eigenvalues. D Therefore, in order to estimate d, one may count the number z of zero eigenvalues of M and choose any d < z. Within this range, smaller values of d will yield more compact representations, while larger values of d will yield more expressive ones, i.e. ones that are most faithful to the original data. What happens in non-ideal conditions, i.e. when the data are not exactly locally fiat , and when one has to contend with numerical noise? The appendix provides an argument showing that the statement in the proposition is robust with respect to ,,' ,, ' ,, ' ,, ' ,, ' ,,' ,,' ,, ' ,,' ,, ' ,, ' 10 ' 0 10" ,,' 2nd e igen value 10" 2nd c igc nvul uc 10 " 10 " 10" lst eige nva] ue 10" 10" 0 0 .,lst c i'cnv? "' " Figure 3: (Left) Eigenvalues for the straight-line data Xi used for Figure 2. (Right) Eigenvalues for the curve data shown in the top-left panel of Figure 1. In both cases the two last eigenvalue are orders of magnitude smaller than the other eigenvalues, indicating a maximal dimension d = 1 for the data. noise, i.e. numerical errors and small deviations from the ideal locally flat data will result in small deviations from the ideal zero-value of the first d + 1 eigenvalues, where d is used here for the 'intrinsic' dimension of the data. This is illustrated in Figure 3. In Figure 4 we describe the successful application of the dimensionality detection method on a data set of synthetically generated grayscale images. 4 LLE and grouping In the first example (2.1) we pointed out the limits of LLE when applied to multiple components of data. It appears then that a grouping procedure should always preceed LLE. The data would be first split into its component groups, each one of which should be then analyzed with LLE. A deeper analysis of the algorithm though, suggests that grouping and LLE could actually be performed at the same time. Proposition 2. Suppose the data set {Xdi=l ,... ,N E ll~P is partitioned into m Kconnected components. Then there exists an m-dimensional eigenspace of M with zero eigenvalue which admits a basis {vih=l,... ,m where the Vi have entries that are either '1' or '0' . More precisely: each Vi corresponds to one of the groups of the data and takes value V i ,j = 1 for j in the group, V i ,j = 0 for j not in the group. Proof. Without loss of generality, assume that the indexing of the data X i is such that the weight matrix W , and consequent ely the matrix M, are block-diagonal with m blocks, each block corresponding to one of the groups of data. This is achieved by a permutation of indices, which will not effect any further step of our algorithm. As a direct consequence of the row normalization of W, each block of M has exactly one eigenvector composed of all ones, with eigenvalue O. Therefore, there is an m-dimensional eigenspace with eigenvalue 0, and there exist a basis of it, each vector of which has value 1 on a certain component, 0 otherwise. D Therefore one may count the number of connected components by computing the eigenvectors of M corresponding to eigenvalue 0, and counting the number m of those vectors Vi whose components take few discrete values (see Figure 6). Each index i may be assigned to a group by clustering based on the value of Vl, ... , V m . Figure 4: (Left) A sample from a data set of N=1000, 40 by 40 grayscale images, each one thought as a point in a 1600 dimensional vector space. In each image, a slightly blurred line separates a dark from a bright portion. The orientation of the line and its distance from the center of the image are variable. (Middle) The non-zero eigenvalues of M. LLE is performed with K=20. The 2nd and 3rd smallest eigenvalues are of smaller size than the others, giving an upper bound of 2 on the intrinsic dimension of the data set. (Right) The 2-dimensional LLE representation. The polar coordinates, after rescaling, are the distance of the dividing line from the center and its orientation. ". Figure 5: The data set is analogous to the one used above (N =1000, 40 by 40 grayscale images, LLE performed with K=20). The orientation of the line dividing the dark from the bright portion is now only allowed to vary in two disjoint intervals. (Middle) The non-zero eigenvalues of M. (Left and Right) The 3rd and 5th (resp. 4th and 6th) eigenvectors of M are used for the LLE representation of the first (resp. the second) K-component. ,,' ,, ' ,, ' ,,' ,, ' 10'0 4th. 5th and 6th eigenvalues 10 " 10 " 10 " 1st. 2nd and 3rd eigenvalues Figure 6: (Left) The last six eigenvectors of M for the broken parabola of Figure 1 shown, top to bottom, in reverse order of magnitude of the corresponding eigenvalue. The x axis is associated to the index i. (Right) The eigenvalues of the same (log scale). Notice that the last six are practically zero. The eigenvectors corresponding to the three last eigenvalues have discrete values indicating that the data is split in three groups. There are z =6 zero-eigenvalues indicating that the dimension of the data is d:::; z/m - 1 = 1. In the Appendix (A) we show that such a process is robust with respect to numerical noise. It is also robust to small perturbations of the block-diagonal structure of M (see Figure 7). This makes the use of LLE for grouping purposes convenient. Should the K-connected components be completely separated, the partition would be easily obtained via a more efficient graph-search algorithm. The proof is carried out for ordered indices as in Fig. 3 but it is invariant under index permutation. The analysis of Proposition 1 may be extended to the dimension of each of the m groups according to Proposition 2. Therefore, in the ideal case, we will find z zero-eigenvalues of M which, together with the number m obtained by counting the discrete-valued eigenvectors may be used to estimate the maximal d using z ~ m(d + 1). This behavior may be observed experimentally, see Figures 6 and 5. 5 Conclusions We have examined two difficulties of the Locally Linear Embedding method [2] and shown that, in a neighborhood of ideal conditions, they may be solved by a careful exam of eigenvectors of the matrix M that are associated to very small eigenvalues. More specifically: the number of groups in which the data is partitioned corresponds to the number of discrete-valued eigenvectors, while the maximal dimension d of the low-dimensional embedding may be obtained by dividing the number of small eigenvalues by m and subtracting 1. Both the groups and the low-dimensional embedding coordinates may be computed from the components of such eigenvectors. Our algorithms have mainly been tested on synthetically generated data. Further investigation on real data sets is necessary in order to validate our theoretical results. w' ,----~-~-~-~--~______, w' w' w' w' w' w' 3rd and 4th eigenvalues . 10 '? 10 " 10 " 10'" 1st and 2ndeige!\values o~--=-----::------=-------=------,=--~ Figure 7: (Left) 2D Data Xi distributed along a broken parabola. Nevertheless, for K=14, the components are not completely K-disconnected (a different symbol is used for the neighbors of the leftmost point on the rightmost component). (Right) The set of eigenvalues for M. A set of two almost-zero eigenvalues and a set of two of small size are visible. References [1] C. Bishop, Neural Networks for Pattern Recognition, Oxford Univ. Press, (1995). [2] S. T. Roweis, L.K.Saul, Science, 290, p. 2323-2326, (2000). [3] J. Tenenbaum , V. de Silva, J. Langford, Science, 290, p. 2319-2323, (2000). A Appendix In Proposition 2 of Section 4 we proved that during the LLE procedure we can automatically detect the number of K -connected components, in case there is no noise. Similarly, in Proposition 1 of Section 3 we proved that under ideal conditions (no noise, locally flat data), we can determine an estimate for the intrinsic dimension of the data. Our next goal is to establish a certain robustness of these results in the case there is numerical noise, or the components are not completely separated, or the data is not exactly locally flat . In general, suppose we have a non degenerate matrix A, and an orthonormal basis of eigenvectors VI, ... , V m , with eigenvalues AI , ... Am. As a consequence of a small perturbation of the matrix into A + dA, we will have eigenvectors Vi + dVi with eigenvalues Ai + dAi' The unitary norm constraint makes sure that dVi is orthogonal to Vi and could be therefore written as dVi = L:k#i O'.ikVk. Using again the orthonormality, one can derive expressions for the perturbations of Ai and Vi : dAi O'.ij (Ai - Aj) < vi,dAvi > < Vj,dAVi > . This shows that if the perturbation dA has order E, then the perturbations dA and are also of order E. Notice that we are not interested in perturbations O'.ij within the eigenspace of eigenvalue 0, but rather those orthogonal to it, and therefore Ai =j:. Aj. O'.ij
2033 |@word middle:2 norm:1 nd:4 covariance:4 pick:1 yih:1 tr:1 carry:1 reduction:2 contains:1 rightmost:1 written:1 finest:1 visible:1 numerical:4 partition:6 shape:1 designed:1 davi:2 fewer:1 parametrization:2 provides:2 along:5 direct:1 consists:1 prove:1 behavior:1 globally:1 automatically:3 estimating:1 moreover:1 panel:1 eigenspace:3 what:3 marzia:1 eigenvector:3 astronomy:1 every:4 xd:1 gravity:3 exactly:3 wrong:2 unit:2 overestimate:1 local:3 limit:1 consequence:2 oxford:1 approximately:1 chose:1 examined:1 suggests:1 range:1 faithful:1 block:5 procedure:2 thought:1 convenient:1 convenience:1 close:3 imposed:1 center:5 yt:1 insight:2 orthonormal:1 embedding:9 handle:2 notion:1 coordinate:6 variation:1 analogous:1 resp:2 target:1 construction:1 suppose:3 ixi:1 hypothesis:1 origin:2 approximated:1 recognition:1 parabola:2 bottom:3 observed:1 solved:2 calculate:2 connected:10 principled:1 broken:2 ui:1 division:2 completely:4 basis:3 easily:2 lst:2 engeneering:1 separated:3 univ:1 describe:1 neighborhood:1 whose:2 encoded:2 larger:1 valued:2 say:1 otherwise:1 itself:1 sequence:1 eigenvalue:34 subtracting:1 maximal:3 degenerate:1 roweis:5 intuitive:1 validate:1 recipe:1 derive:1 exam:1 nearest:3 ij:4 dividing:3 correct:1 investigation:1 proposition:7 cnv:1 practically:1 vary:1 consecutive:1 smallest:2 purpose:1 uniqueness:1 polar:1 lose:1 faithfully:1 clearly:1 always:1 rather:2 rank:2 mainly:1 wf:1 detect:1 am:1 vl:1 pasadena:2 perona:2 wij:3 interested:1 orientation:3 priori:1 nva:1 uc:1 equal:2 having:1 placing:1 represents:2 others:1 few:1 composed:3 simultaneously:1 cheaper:1 detection:2 analyzed:1 necessary:1 orthogonal:3 desired:1 theoretical:1 column:2 cost:2 deviation:2 entry:1 successful:1 xdi:1 answer:1 my:1 st:2 physic:1 together:1 again:1 choose:2 possibly:1 preceed:1 rescaling:1 suggesting:1 de:1 coefficient:2 blurred:1 explicitly:1 ely:1 vi:8 performed:3 closed:1 portion:2 square:1 bright:2 yield:3 produced:1 finer:1 straight:3 addend:1 associated:3 proof:3 proved:2 intrinsically:1 dimensionality:5 fiat:2 actually:1 appears:1 done:1 though:1 generality:1 furthermore:1 langford:1 expressive:1 nonlinear:1 aj:2 indicated:1 effect:2 normalized:1 true:1 vih:1 orthonormality:1 hence:2 assigned:1 illustrated:2 ll:1 during:1 ue:1 chosed:1 criterion:1 leftmost:1 tt:1 silva:1 image:5 recently:2 common:2 behaves:1 endpoint:1 polito:2 belong:1 approximates:1 onedimensional:1 ai:5 rd:11 mathematics:2 similarly:1 pointed:1 iyi:1 longer:1 curvature:1 discard:1 reverse:1 store:1 certain:2 yi:14 caltech:2 minimum:1 dai:2 somewhat:1 wijxj:2 determine:3 multiple:1 divided:1 manipulate:1 variant:1 circumstance:1 kernel:2 normalization:1 achieved:1 interval:1 singular:1 sure:1 elegant:2 integer:1 call:1 unitary:1 counting:2 ideal:6 synthetically:2 vital:1 split:2 xj:1 perfectly:1 six:2 pca:2 expression:1 eigenvectors:12 dark:2 locally:12 tenenbaum:1 inspiring:1 exist:1 notice:3 estimated:1 disjoint:2 correctly:1 discrete:4 shall:1 group:19 key:1 nevertheless:1 drawn:1 graph:1 pietro:1 almost:1 appendix:3 bound:3 q9:1 quadratic:1 constraint:7 precisely:1 flat:4 generates:1 argument:1 according:1 combination:4 disconnected:3 belonging:1 beneficial:1 smaller:3 slightly:1 partitioned:2 happens:2 invariant:2 indexing:1 r3:2 count:2 know:1 robustness:1 original:1 top:4 clustering:1 ensure:2 recognizes:1 igc:1 calculating:1 giving:1 uj:1 establish:1 classical:1 diagonal:2 pathologically:1 subspace:2 distance:2 separate:2 manifold:6 enforcing:1 index:8 minimizing:3 statement:1 contend:1 upper:3 discarded:1 curved:1 situation:1 extended:1 perturbation:6 reproducing:1 arbitrary:1 introduced:1 california:2 pattern:1 difficulty:1 technology:2 igen:1 axis:4 carried:1 embedded:1 loss:1 permutation:2 affine:1 dvi:3 share:1 translation:2 row:2 compatible:1 last:4 lle:25 deeper:1 institute:2 saul:5 neighbor:11 sparse:1 distributed:5 curve:5 dimension:14 calculated:3 collection:1 compact:1 dealing:1 xi:11 grayscale:3 search:1 robust:3 ca:2 constructing:1 da:3 vj:1 whole:1 noise:6 allowed:1 fig:1 fails:1 lie:4 bishop:1 showing:1 symbol:2 admits:1 consequent:1 grouping:5 intrinsic:6 exists:2 magnitude:2 visual:1 contained:1 ordered:2 corresponds:3 goal:1 careful:1 experimentally:1 specifically:1 principal:1 invariance:1 indicating:3 tested:1
1,134
2,034
Receptive field structure of flow detectors for heading perception Jaap A. B e intema Dept. Zoology & Neurobiology Ruhr University Bochum, Germany, 44780 beintema@neurobiologie.ruhr-uni-bochum.de Albert V. van den Berg Dept. of Neuro-ethology, Helmholtz Institute, Utrecht University, The Netherlands a. v. vandenberg@bio.uu.nl Markus Lappe Dept. Zoology & Neurobiology Ruhr University Bochum, Germany, 44780 lappe@neurobiologie .ruhr-uni-bochum.de Abstract Observer translation relative to the world creates image flow that expands from the observer's direction of translation (heading) from which the observer can recover heading direction. Yet, the image flow is often more complex, depending on rotation of the eye, scene layout and translation velocity. A number of models [1-4] have been proposed on how the human visual system extracts heading from flow in a neurophysiologic ally plausible way. These models represent heading by a set of neurons that respond to large image flow patterns and receive input from motion sensed at different image locations. We analysed these models to determine the exact receptive field of these heading detectors. We find most models predict that, contrary to widespread believe, the contribut ing motion sensors have a preferred motion directed circularly rather than radially around the detector's preferred heading. Moreover, the results suggest to look for more refined structure within the circular flow, such as bi-circularity or local motion-opponency. Introduction The image flow can be considerably more complicated than merely an expanding pattern of motion vectors centered on the heading direction (Fig. 1). Flow caused by eye rotation (Fig. 1b) causes the center of flow to be displaced (compare Fig. 1a and c). The effect of rotation depends on the ratio ofrotation and translation speed. A Translational flow f~~~~; ,,~ ~ t // ...... ........- 1 ?0 ? " II .J + ? J ..... ? ?.,. "t B <If:- .... --- ... ~ ~~~ S:: ... ... ....-.... +-....... ~ Combined flow ........ . - ., .... '\ - 4i- .. .... ..+-..,... ....,.. . .. * :: +- ~1 " .. ...... --... ',,,, ...... \\ .. 1J: ~ ... ............... -+ -+~ C Rotational flow +- ....... ... 0 '# +- ~ I' ~ ...... ... 1 ~.. " ? ",. ~ 4 ~\ + ~ .1, Figure 1: Flow during a) observer translation through a 3D-cloud of dots, headed 10? towards the left, during b) observer rotation about the vertical towards the right, and during c) the combination of both. Also, since the image motions caused by translation depend on point distance and the image motions caused by rotation do not, the combined movement results in flow that is no longer purely expanding for scenes containing depth differences (Fig. lc). Heading detection can therefore not rely on a simple extrapolation mechanism that determines the point of intersection of motion vectors. A number of physiologically-based models [1-4] have been proposed on how the visual system might arrive at a representation of heading from flow that is insensit ive to parameters other than heading direction. These models assume heading is encoded by a set of units that each respond best to a specific pattern of flow that matches their preferred heading. Such units resemble neurons found in monkey brain area MST. MST cells have large receptive fields (RF), typically covering one quart or more of the visual field, and receive input from several local motion sensors in brain area MT. The receptive field of MST neurons may thus be defined as the preferred location, speed and direction of all input local motion sensors. Little is known yet about the RF structure of MST neurons. We looked for similarities between current models at the level of the RF structure. First we explain the RF structure of units in the velocity gain model, because this model makes clear assumptions on the RF structure. Next, we we show the results of reconstructing RF structure of units in the population model[2] . Finally, we analyse the RF structure of the template model[3] and motion-opponency model[4]. Velocity gain field model The velocity gain field model[l] is based on flow templates. A flow template, as introduced by Perrone and Stone[3] , is a unit that evaluates the evidence that the flow fits the unit 's preferred flow field by summing the responses of local motion sensors outputs. Heading is then represented by the preferred heading direction of the most active template(s) . The velocity gain field model[l] is different from Perrone and Stone's template model[2] in the way it acquires invariance for translation speed, point distances and eye rotation. Whereas the template model requires a different template for each possible combination of heading direction and rotation, t he velocity gain field model obtains rotation invariance using far less templates by exploiting eye rotation velocity signals. The general scheme applied in the velocity gain field model is as follows. In a set of flow templates, each tuned to pure expansion with specific preferred heading, B A Circular component <i!t- -. l .... ... t ~ 4t"44- ~ .... 9' ~ ...0 ? .... ............ 9'9'? II ,9' "" +-4- II #= . ~ II ?s- " ? II ~ ~ + 9'Y 9' ? ??? .. Radial component ,\ ~ ~~ .t. ? 1'f ;J? "" .......... r .J' t ,-' Jr y y 0 .. " ,//, J ? ... ... -+ ?? \ ~. --.- .~ -...; .. ~-. .. Figure 2: The heading-centered circular (a) and radial (b) component of the flow during combined translation and rotation as in Fig. 2c. the templates would change their activity during eye rotation. Simply subtracting the rotation velocity signal for each flow template would not suffice to compensate because each template is differently affected by rotational flow. However, each flow template can become approximately rotation-invariant by subtracting a gain field activity that is a multiplication of the eye velocity t with a derivative template activity 80/ 8R that is specific for each flow template. The latter reflects the change in flow template activity 0 given a change in rotational flow 8R. Such derivative template 80/ 8R can be constructed from the activity difference of two templates tuned to the same heading, but opposite rotation. Thus, in the velocity gain field model, templates tuned to heading direction and a component of rotation play an important role. To further appreciate the idea behind the RF structure in the velocity gain field model, note that the retinal flow can be split into a circular and radial component, centered on the heading point (Fig. 2). Translation at different speeds or through a different 3D environment will alter the radial component only. The circular component contains a rotational component of flow but does not change with point distances or translational speed. This observation lead to the assumption implemented in the velocity gain field model that templates should only measure the flow along circles centered on the point of preferred heading. An example of the RF structure of a typical unit in the velocity gain field model, tuned to heading and rightward rotation is shown in Fig. 3. This circular RF structure strongly reduces sensitivity to variations in depth structure or the translational speed, while the template's tuning to heading direction is preserved, because its preferred structure is centered on its preferred heading direction [1] . Interestingly, the RF structure of the typical rotation-tuned heading units is bi-circular, because the direction of circular flow is opponent in the hemifields to either side of an axis (in this case the horizontal axis) through the heading point. Moreover, the structure contains a gradient in magnitude along the circle, decreasing towards the horizontal axis. F:- . ~ . ..... ... .r- I\. ..... I\. 1\ ? ? ? ? I\. It 0 I "'- . - " .? .... -- ... " It ..... .... .r Of' ~ ? .. rJ Figure 3: Bi-circular RF structure of a typical unit in the velocity gain field model, tuned to leftward heading and simultaneous rightward rotation about the vertical. Individual vectors show the preferred direction and velocity of the input motion sensors. Population model The population model [2] derives a representation of heading direction that is invariant to the other flow parameters using a totally different approach. This model does not presume an explicit RF structure. Instead, the connections strengths and preferred directions of local motion inputs to heading-specific flow units are computed according to an optimizing algorithm[5]. We here present the results obtained for a restricted version of the model in which eye rotation is assumed to be limited to pursuit that keeps the eye fixated on a stationary point in the scene during the observer translation. Specifically, we investigated whether a circular or bi-circular RF structure as predicted by the velocity gain model emerges in the population model. The population model [2 ,6] is an implementation of the subspace algorithm by Heeger and Jepson [5] into a neural network. The subspace algorithm computes a residual function R(T j) for a range of possible preferred heading directions. The residual function is minimized when flow vectors measured at m image locations, described as one array, are perpendicular to the vectors t hat form columns of a matrix C ~ (T j). This matrix is computed from the preferred 3-D translation vector T j and the m image locations. Thus, by finding the matrix that minimizes the residue, the algorithm has solved the heading, irrespective of the 3D-rotation vector, unknown depths of points and translation speed. To implement the subspace algorithm in a neurophysiologically plausible way, the population model assumes two layers of units. The first MT-like layer contains local motion sensors that fire linearly with speed and have cosine-like direction tuning. These sensors connect to units in the second MST-like layer. The activity in a 2nd layer unit , with specific preferred heading T j, represents the likelihood that the residual function is zero. The connection strengths are determined by the C ~ (T j) matrix. As not to have too many motion inputs per 2nd layer unit, the residual function R(T j) is partitioned into smaller sub residues that take only a few motion inputs. The likelihood for a specific heading is t hen given by the sum of responses in a population with same preferred heading. Given the image locations and the preferred heading, one can reconstruct the RF structure for 2nd layer units with the same preferred heading. The preferred motion inputs to a second layer unit are given by vectors t hat make up each column of C ~ (T j). Hereby, t he vector direction represents the preferred motion direction, A B "- \ t I ? Figure 4: Examples of receptive field structure of a population that encodes heading 100 towards the left (circle) . a-b) Five pairs of MT-like sensors, where the motion sensors of each pair are at a) the same image location, or b) at image locations one quarter of a cycle apart. c) Distribution of multiple pairs leading to bi-circular pattern. and the vector magnitude represents the strength of the synaptic connection. The matrix C l..(Tj) is computed from the orthogonal complement of a (2m x m + 3) matrix C(Tj) [5]. On the assumption that only fixational eye movements occur, the matrix reduces to (2m x m + 1)[6]. Given only two flow vector inputs (m = 2), the matrix C l.. (T j) reduces to one column of length m = 4. The orthogonal complement of this 4 x 3 matrix was solved in Mathematica by first computing the nullspace of the inverse matrix of C (T j), and then constructing an orthonormal basis for it using Gram-Schmidt orthogonalisation. We computed the orientation and magnitude of the two MT-inputs analytically. Instead of giving the mathematics, we here describe the main results. Circularity Independent of the spatial arrangement of the two MT-inputs to a 2nd-layer unit, their preferred motions turned out to be always directed along a circle centered on the preferred heading point. Fig. 4 shows examples of the circular RF structures, for different distributions of motion pairs that code for the same heading direction. Motion-opponency For pairs of motion sensors at overlapping locations, the vectors of each pair always turned out to be opponent and of equal magnitude (Fig. 4a). For pairs of spatially separated motion sensors, the preferred magnitude and direction of the two motion inputs depend on their location with respect to the hemispheres divided by the line through heading and fixation point. We find that preferred motion directions are opponent if the pair is located within the same hemifield, but uni-directional if the pair is split across the two hemifields as in Fig. 4b. Bi-circularity Interestingly, if pairs of motion sensors are split across hemi fields, with partners at image locations 90 0 rotated about the heading point, a magnitude gradient appears in the RF structure (Fig. 4b). Thus, with these pairs a bi-circular RF structure can be constructed similar to units tuned to rotation about the vertical in the velocity gain field model (compare with Fig. 3). Note, that the bi-circular RF structures do differ since the axis along which the largest magnitude occurs is horizontal for the population model and vertical for the velocity gain field model. The RF structure of the population model unit resembles a velocity gain field unit tuned to rotation about the horizontal axis, implying a Adapted from Perrone and Stone (1994) A Effective RF structure B Direction and speed tuned motion sensors ""","" .,"',,", ~ (j Figure 5: Adapted from Perrone and Stone 1994). a) Each detector sums the responses of the most active sensor at each location. This most active motion sensor is selected from a pool of sensors tuned to different depth planes (Ca, Cb, etc). These vectors are the vector sums of preferred rotation component Rand translational components Ta, Tb, etc. b) Effective RF structure. large sensitivity to such rotation. This, however, does not conflict with the expected performance of the population model. Because in this restricted version rotation invariance is expected only for rotation that keeps the point of interest in the center of the image plane (in this case rotation about the vertical because heading is leftward) units are likely to be sensitive to rotation about the horizontal and torsional axis. Template model The template model and the velocity gain field model differ in how invariance for translation velocities, depth structure and eye rotation is obtained. Here, we investigate whether this difference affects the predicted RF structure. In the template model of Perrone and Stone [3], a template invariant to translation velocity or depth structure is obtained by summing the responses of the most active sensor at each image location. This most active sensor is selected from a collection of motion sensors, each tuned to a different ego-translation speed (or depth plane), but with the same preferred ego-rotation and heading direction (Fig. 5a). Given a large range of depth planes, it follows that a different radial component of motion will stimulate another sensor maximally, but that activity nevertheless remains the same. The contributing response will change only due to a component of motion along a circle centered on the heading, such as is the case when heading direction or rotation is varied. Thus, the contributing response will always be from the motion sensor oriented along the circle around the template's preferred heading. Effectively, this leads to a bi-circular RF structure for units tuned to heading and rotation (Fig. 5b). Motion-opponency model Royden[4] proposed that the effect of rotation is removed at local motion detection level before the motion signals are received by flow detectors. This is achieved by MT-like sensors that compute the difference vector between spatially neighbouring motion vectors. Such difference vector will always be oriented along lines intersecting at the heading point (Fig. 6). Thus, the resulting input to flow detectors will be oriented radially. Indeed, Royden's results[4] show that the preferred directions of the operators with the largest response will be radially, not circularly, oriented. A Translational flow B Rotational flow L -________ ~ . ~ ______ ~ Figure 6: Motion parallax, the difference vector between locally neighbouring motion vectors. For translation flow (a) the difference vector will be oriented along line through the heading point, whereas for rotational flow (b) the difference vector vanishes (compare vectors within square). Summary and Discussion We showed that a circular RF structure, such as proposed by the velocity gain field model[l] , is also found in the population model[2] and is effectively present in the template model[3] as well. Only the motion-opponent model [4] prefers radial RF structures. Furthermore, we find that under certain restrictions, the population model reveals local motion-opponency and bi-circularity, properties that can be found in the other models as well. A circular RF structure turns out to be a prominent property in three models. This supports the counterintuitive, but computationally sensible idea, that it is not the radial flow structure, but the structure perpendicular to it, that contributes to the response of heading-sensitive units in the human brain. Studies on area MST cells not only report selectivity for expanding motion patterns, but also a significant proportion of cells that are selective to rotation patterns [7-10]. These models could explain why cells respond so well to circular motion, in particular to the high rotation speeds (up to about 80 deg/s) not experienced in daily life. This model study suggests that selectivity for circular flow has a direct link to heading detection mechanisms. It also suggests that testing selectivity for expanding motion might be a bad indicator for determining a cell's preferred heading. This point has been noted before, as MST seems to be systematically tuned to the focus of rotation, exactly like model neurons [9]. Little is still known about the receptive field structure of MST cells. So far the receptive field structure of MST cells has only been roughly probed [10], and the results neither support a radial nor a circular structure. Also , so far only uni-circular motion has been tested. Our analyses points out that it would be worthwhile to look for more refined circular structure such as local motion-opponency. Local motion opponency has already been found in area MT, where some cells respond only if different parts of their receptive field are stimulated with different motion [11]. Another promising structure to look for would be bi-circularity, with gradients in magnitude of preferred motion along the circles. Acknowledgments Supported by the German Science Foundation and the German Federal Ministry of Education and Research. References [1] Beintema, J . A. & van den Berg A. V. (1998) Heading detection using motion templates and eye velocity gain fields. Vision Research, 38(14):2155-2179. [2] Lappe M., & Rauschecker J . P. (1993) A neural network for the processing of optic flow from ego-motion in man and higher mammals. Neural Computation, 5:374-39l. [3] Perrone J. A. & Stone L. S. (1994) A model for the self-motion estimation within primate extrastriate visual cortex. Vision Research, 34:2917-2938 . [4] Royden C. S. (1997) Mathematical analysis of motion-opponent mechanisms used in the determination of heading and depth. Journal of th e Optical Society of America A, 14(9):2128-2143. [5] Heeger D . J . & Jepson A. D . (1992) Subspace methods for recovering rigid motion I: Algorithm and implementation. International Journal of Computational Vision , 7:95-117. [6] Lappe M. & Rauschecker J.P. (1993) Computation of heading direction from optic flow in visual cortex. In C.L. Giles, S.J. Hanson and J.D. Cowan (eds.), Advances in Neural Information Processing Systems 5, pp. 433-440. Morgan Kaufmann. [7] Tanaka K. & Saito H. (1989) Analysis of the visual field by direction, expansion/contraction, and rotation cells clustered in the dorsal part of the medial superior temporal area of the macaque monkey Journal of Neurophysiology, 62(3):626-64l. [8] Duffy C. J. & Wurtz R. H. (1991) Sensitivity of MST neurons to optic flow stimuli. I. A continuum of response selectivity to large-field stimuli. Journal of Neurophysiology, 65(6) :1329-1345. [9] Lappe M., Bremmer F ., Pekel M., Thiele A., Hoffmann K.-P. (1996) Optic flow processing in monkey STS: a theoretical and experimental approach . th e Journal of Neuroscience, 16(19):6265-6285. [10] Duffy C. J. & Wurtz R. H. (1991) Sensitivity of MST neurons to optic flow stimuli. II. Mechanisms of response selectivity revealed by small-field stimuli. Journal of Neurophysiology, 65(6):1346-1359. [11] Allman J., Miezin F. & McGuinness E. (1985) Stimulus specific responses from beyond the classical receptive field: Neurophysiological mechanisms for local-global comparisons in visual neurons. Ann. R ev. N eurosci., 8:407-430.
2034 |@word neurophysiology:3 version:2 proportion:1 seems:1 nd:4 ruhr:4 sensed:1 contraction:1 mammal:1 extrastriate:1 contains:3 tuned:13 interestingly:2 current:1 analysed:1 yet:2 mst:11 medial:1 stationary:1 implying:1 selected:2 plane:4 location:12 five:1 mathematical:1 along:9 constructed:2 direct:1 become:1 fixation:1 headed:1 parallax:1 indeed:1 expected:2 roughly:1 nor:1 brain:3 decreasing:1 little:2 totally:1 moreover:2 suffice:1 minimizes:1 monkey:3 finding:1 temporal:1 expands:1 exactly:1 bio:1 unit:23 before:2 local:11 torsional:1 approximately:1 might:2 resembles:1 suggests:2 limited:1 bi:11 range:2 perpendicular:2 directed:2 acknowledgment:1 testing:1 implement:1 saito:1 area:5 radial:8 suggest:1 bochum:4 operator:1 hemi:1 restriction:1 center:2 layout:1 pure:1 array:1 orthonormal:1 counterintuitive:1 population:13 variation:1 play:1 exact:1 neighbouring:2 velocity:25 helmholtz:1 ego:3 located:1 cloud:1 role:1 solved:2 cycle:1 movement:2 removed:1 thiele:1 environment:1 vanishes:1 depend:2 purely:1 creates:1 basis:1 rightward:2 differently:1 represented:1 america:1 separated:1 describe:1 effective:2 refined:2 encoded:1 plausible:2 ive:1 reconstruct:1 analyse:1 orthogonalisation:1 subtracting:2 turned:2 beintema:2 exploiting:1 rotated:1 depending:1 measured:1 received:1 implemented:1 predicted:2 resemble:1 recovering:1 uu:1 differ:2 direction:27 centered:7 human:2 education:1 clustered:1 around:2 cb:1 predict:1 continuum:1 eurosci:1 estimation:1 sensitive:2 largest:2 reflects:1 federal:1 sensor:22 always:4 rather:1 focus:1 likelihood:2 rigid:1 typically:1 selective:1 germany:2 translational:5 orientation:1 spatial:1 field:32 equal:1 represents:3 look:3 alter:1 minimized:1 report:1 stimulus:5 few:1 oriented:5 individual:1 fire:1 detection:4 interest:1 circular:23 investigate:1 zoology:2 nl:1 circularity:5 behind:1 tj:2 daily:1 bremmer:1 orthogonal:2 circle:7 opponency:7 theoretical:1 column:3 giles:1 too:1 connect:1 considerably:1 combined:3 neurobiologie:2 sensitivity:4 international:1 pool:1 hemifield:1 intersecting:1 containing:1 derivative:2 leading:1 de:2 retinal:1 sts:1 caused:3 depends:1 observer:6 extrapolation:1 jaap:1 recover:1 complicated:1 square:1 kaufmann:1 directional:1 utrecht:1 presume:1 detector:6 explain:2 simultaneous:1 synaptic:1 ed:1 evaluates:1 mathematica:1 pp:1 hereby:1 gain:19 radially:3 emerges:1 appears:1 ta:1 higher:1 response:11 maximally:1 rand:1 strongly:1 furthermore:1 ally:1 horizontal:5 overlapping:1 widespread:1 stimulate:1 believe:1 effect:2 analytically:1 spatially:2 during:6 self:1 covering:1 acquires:1 noted:1 cosine:1 prominent:1 stone:6 motion:55 image:15 superior:1 rotation:37 quarter:1 mt:7 he:2 significant:1 tuning:2 mathematics:1 dot:1 longer:1 similarity:1 cortex:2 etc:2 showed:1 leftward:2 optimizing:1 hemisphere:1 apart:1 selectivity:5 certain:1 life:1 morgan:1 ministry:1 determine:1 signal:3 ii:6 multiple:1 rj:1 reduces:3 ing:1 match:1 determination:1 compensate:1 divided:1 neuro:1 vision:3 wurtz:2 albert:1 represent:1 achieved:1 cell:9 receive:2 whereas:2 preserved:1 residue:2 cowan:1 contrary:1 flow:49 allman:1 revealed:1 split:3 affect:1 fit:1 opposite:1 idea:2 lappe:5 whether:2 cause:1 prefers:1 clear:1 quart:1 netherlands:1 locally:1 fixational:1 neuroscience:1 per:1 probed:1 affected:1 nevertheless:1 neither:1 mcguinness:1 merely:1 sum:3 inverse:1 respond:4 arrive:1 layer:8 activity:7 strength:3 occur:1 adapted:2 optic:5 scene:3 encodes:1 markus:1 speed:11 optical:1 hemifields:2 according:1 combination:2 perrone:6 jr:1 smaller:1 across:2 reconstructing:1 partitioned:1 primate:1 den:2 invariant:3 restricted:2 computationally:1 remains:1 turn:1 german:2 mechanism:5 pursuit:1 opponent:5 worthwhile:1 schmidt:1 hat:2 assumes:1 giving:1 society:1 classical:1 appreciate:1 arrangement:1 already:1 looked:1 occurs:1 receptive:9 hoffmann:1 gradient:3 subspace:4 distance:3 link:1 sensible:1 partner:1 length:1 code:1 ratio:1 rotational:6 implementation:2 unknown:1 vertical:5 neuron:8 displaced:1 observation:1 neurobiology:2 varied:1 introduced:1 complement:2 pair:11 connection:3 rauschecker:2 conflict:1 hanson:1 tanaka:1 macaque:1 beyond:1 perception:1 pattern:6 ev:1 tb:1 rf:27 rely:1 indicator:1 residual:4 scheme:1 eye:11 ethology:1 axis:6 irrespective:1 extract:1 hen:1 multiplication:1 contributing:2 relative:1 determining:1 neurophysiologically:1 foundation:1 systematically:1 translation:16 summary:1 supported:1 heading:56 side:1 institute:1 template:28 van:2 depth:9 world:1 gram:1 computes:1 collection:1 far:3 obtains:1 uni:4 preferred:30 keep:2 deg:1 global:1 active:5 reveals:1 fixated:1 summing:2 assumed:1 physiologically:1 why:1 stimulated:1 promising:1 expanding:4 ca:1 contributes:1 expansion:2 investigated:1 complex:1 constructing:1 jepson:2 main:1 linearly:1 fig:15 lc:1 sub:1 experienced:1 explicit:1 heeger:2 nullspace:1 miezin:1 bad:1 specific:7 evidence:1 derives:1 circularly:2 effectively:2 magnitude:8 duffy:2 intersection:1 simply:1 likely:1 neurophysiological:1 visual:7 determines:1 ann:1 towards:4 man:1 change:5 typical:3 specifically:1 determined:1 invariance:4 experimental:1 berg:2 support:2 latter:1 dorsal:1 dept:3 tested:1
1,135
2,035
A Bayesian Network for Real-Time Musical Accompaniment Christopher Raphael Department of Mathematics and Statistics, University of Massachusetts at Amherst, Amherst, MA 01003-4515, raphael~math.umass.edu Abstract We describe a computer system that provides a real-time musical accompaniment for a live soloist in a piece of non-improvised music for soloist and accompaniment. A Bayesian network is developed that represents the joint distribution on the times at which the solo and accompaniment notes are played, relating the two parts through a layer of hidden variables. The network is first constructed using the rhythmic information contained in the musical score. The network is then trained to capture the musical interpretations of the soloist and accompanist in an off-line rehearsal phase. During live accompaniment the learned distribution of the network is combined with a real-time analysis of the soloist's acoustic signal, performed with a hidden Markov model, to generate a musically principled accompaniment that respects all available sources of knowledge. A live demonstration will be provided. 1 Introduction We discuss our continuing work in developing a computer system that plays the role of a musical accompanist in a piece of non-improvisatory music for soloist and accompaniment. The system begins with the musical score to a given piece of music. Then, using training for the accompaniment part as well as a series of rehearsals, we learn a performer-specific model for the rhythmic interpretation of the composition. In performance, the system takes the acoustic signal of the live player and generates the accompaniment around this signal, in real-time, while respecting the learned model and the constraints imposed by the score. The accompaniment played by our system responds both flexibly and expressively to the soloist's musical interpretation. Our system is composed of two high level tasks we call "Listen" and "Play." Listen takes as input the acoustic signal of the soloist and, using a hidden Markov model, performs a real-time analysis of the signal. The output of Listen is essentially a running commentary on the acoustic input which identifies note boundaries in the solo part and communicates these events with variable latency. The HMM framework is well-suited to the listening task and has several attributes we regard as indispensable to any workable solution: 1. The HMM allows unsupervised training using the Baum-Welch algorithm. Thus we can automatically adapt to changes in solo instrument, microphone placement, ambient noise, room acoustics, and the sound of the accompaniment instrument. 2. Musical accompaniment is inherently a real-time problem. Fast dynamic programming algorithms provide the computational efficiency necessary to process the soloist's acoustic signal at a rate consistent with the real-time demands of our application. 3. Musical signals are occasionally ambiguous locally in time, but become easier to parse when more context is considered. Our system owes much of its accuracy to the probabilistic formulation of the HMM. This formulation allows one to compute the probability that an event is in the past. We delay the estimation of the precise location of an event until we are reasonably confident that it is, in fact, past. In this way our system achieves accuracy while retaining the lowest latency possible in the identification of musical events. Our work on the Listen component is documented thoroughly in [1] and we omit a more detailed discussion here. The heart of our system, the Play component, develops a Bayesian network consisting of hundreds of Gaussian random variables including both observable quantities, such as note onset times, and unobservable quantities, such as local tempo. The network can be trained during a rehearsal phase to model both the soloist's and accompanist's interpretations of a specific piece of music. This model then forms the backbone of a principled real-time decision-making engine used in performance. We focus here on the Play component which is the most challenging part of our system. A more detailed treatment of various aspects of this work is given in [2- 4]. 2 Knowledge Sources A musical accompaniment requires the synthesis of a number of different knowledge sources. From a modeling perspective, the fundamental challenge of musical accompaniment is to express these disparate knowledge sources in terms of a common denominator. We describe here the three knowledge sources we use. 1. We work with non-improvisatory music so naturally the musical score, which gives the pitches and relative durations of the various notes, as well as points of synchronization between the soloist and accompaniment, must figure prominently in our model. The score should not be viewed as a rigid grid prescribing the precise times at which musical events will occur; rather, the score gives the basic elastic material which will be stretched in various ways to to produce the actual performance. The score simply does not address most interpretive aspects of performance. 2. Since our accompanist must follow the soloist, the output of the Listen component, which identifies note boundaries in the solo part, constitutes our second knowledge source. While most musical events, such as changes between neighboring diatonic pitches, can be detected very shortly after the change of note, some events, such as rearticulations and octave slurs, are much less obvious and can only be precisely located with the benefit of longer term hindsight. With this in mind, we feel that any successful accompaniment system cannot synchronize in a purely responsive manner. Rather it must be able to predict the future using the past and base its synchronization on these predictions, as human musicians do. 3. While the same player's performance of a particular piece will vary from rendition to rendition, many aspects of musical interpretation are clearly established with only a few repeated examples. These examples, both of solo performances and human (MIDI) performances of the accompaniment part constitute the third knowledge source for our system. The solo data is used primarily to teach the system how to predict the future evolution of the solo part. The accompaniment data is used to learn the musicality necessary to bring the accompaniment to life. We have developed a probabilistic model, a Bayesian network, that represents all of these knowledge sources through a jointly Gaussian distribution containing hundreds of random variables. The observable variables in this model are the estimated soloist note onset times produced by Listen and the directly observable times for the accompaniment notes. Between these observable variables lies a layer of hidden variables that describe unobservable quantities such as local tempo, change in tempo, and rhythmic stress. 3 A Model for Rhythmic Interpretation We begin by describing a model for the sequence of note onset times generated by a monophonic (single voice) musical instrument playing a known piece of music. For each of the notes, indexed by n = 0, . . . , N, we define a random vector representing the time, tn, (in seconds) at which the note begins, and the local "tempo," Sn, (in secs. per measure) for the note. We model this sequence ofrandom vectors through a random difference equation: (1) n = 0, ... , N - 1, where in is the musical length of the nth note, in measures, and the {(Tn' CTnY} and (to, so)t are mutually independent Gaussian random vectors. ? The distributions of the {CT n } will tend concentrate around expressing the notion that tempo changes are gradual. The means and variances of the {CT n} show where the soloist is speeding-up (negative mean), slowing-down (positive mean), and tell us if these tempo changes are nearly deterministic (low variance), or quite variable (high variance). The {Tn} variables describe stretches (positive mean) or compressions (negative mean) in the music that occur without any actual change in tempo, as in a tenuto or agogic accent. The addition of the {Tn} variables leads to a more musically plausible model, since not all variation in note lengths can be explained through tempo variation. Equally important, however, the {Tn} variables stabilize the model by not forcing the model to explain, and hence respond to, all note length variation as tempo variation. Collectively, the distributions of the (Tn' CTn)t vectors characterize the solo player's rhythmic interpretation. Both overall tendencies (means) and the repeatability of these tendencies (covariances) are captured by these distributions. 3.1 Joint Model of Solo and Accompaniment In modeling the situation of musical accompaniment we begin with the our basic rhythm model of Eqn. 1, now applied to the composite rhythm. More precisely, Listen Update Composite Accomp Figure 1: A graphical description of the dependency structure of our model. The top layer of the graph corresponds to the solo note onset times detected by Listen. The 2nd layer of the graph describes the (Tn, 0"n) variables that characterize the rhythmic interpretation. The 3rd layer of the graph is the time-tempo process {(Sn, t n )}. The bottom layer is the observed accompaniment event times. let mo , ... , mivs and mg, ... , m'Na denote the positions, in measures, of the various solo and accompaniment events. For example, a sequence of quarter notes in 3/ 4 time would lie at measure positions 0, 1/ 3, 2/ 3, etc. We then let mo, ... , mN be the sorted union of these two sets of positions with duplicate times removed; thus mo < ml < .. . < mN? We then use the model of Eqn. 1 with In = mn+1 - m n , n = 0, . . . , N - 1. A graphical description of this model is given in the middle two layers of Figure 1. In this figure, the layer labeled "Composite" corresponds to the time-tempo variables, (tn, sn)t, for the composite rhythm, while the layer labeled "Update" corresponds to the interpretation variables (Tn, 0"n) t. The directed arrows of this graph indicate the conditional dependency structure of our model. Thus, given all variables "upstream" of a variable, x, in the graph, the conditional distribution of x depends only on the parent variables. Recall that the Listen component estimates the times at which solo notes begin. How do these estimates figure into our model? We model the note onset times estimated by Listen as noisy observations of the true positions {t n }. Thus if m n is a measure position at which a solo note occurs, then the corresponding estimate from Listen is modeled as an = tn + an 2 where an rv N(O, 1I ). Similarly, if m n is the measure position of an accompaniment event, then we model the observed time at which the event occurs as bn = tn + f3n where f3n rv N(O, ",2). These two collections of observable variables constitute the top layer of our figure, labeled "Listen," and the bottom layer, labeled "Accomp." There are, of course, measure positions at which both solo and accompaniment events should occur. If n indexes such a time then an and bn will both be noisy observations of the true time tn. The vectors/ variables {(to, so)t, (Tn ' O"n)t, a n , f3n} are assumed to be mutually independent. 4 Training the Model Our system learns its rhythmic interpretation by estimating the parameters of the (Tn,O"n) variables. We begin with a collection of J performances of the accompaniment part played in isolation. We refer to the model learned from this accompaniment data as the "practice room" distribution since it reflects the way the accompanist plays when the constraint of following the soloist is absent. For each Listen Update Composite Accomp Figure 2: Conditioning on the observed accompaniment performance (darkened circles), we use the message passing algorithm to compute the conditional distributions on the unobservable {Tn' O"n} variables. such performance, we treat the sequence of times at which accompaniment events occur as observed variables in our model. These variables are shown with darkened circles in Figure 2. Given an initial assignment of of means and covariances to the (Tn , O"n) variables, we use the "message passing" algorithm of Bayesian Networks [8,9] to compute the conditional distributions (given the observed performance) of the (Tn,O"n) variables. Several such performances lead to several such estimates, enabling us to improve our initial estimates by reestimating the (Tn ' O"n) parameters from these conditional distributions. More specifically, we estimate the (Tn,O"n) parameters using the EM algorithm, as follows, as in [7]. We let J-L~, ~~ be our initial mean and covariance matrix for the vector (Tn, 0"n). The conditional distribution of (Tn, 0"n) given the jth accompaniment performance, and using {J-L~ , ~~} , has a N(m; ,n, S~ ) distribution where the m;,n and S~ parameters are computed using the message passing algorithm. We then update our parameter estimates by 1 J . } Lmj,n j=l ~ i+ l n The conventional wisdom of musicians is that the accompaniment should follow the soloist. In past versions of our system we have explicitly modeled the asymmetric roles of soloist and accompaniment through a rather complicated graph structure [2- 4] . At present we deal with this asymmetry in a more ad hoc, however, perhaps more effective, manner , as follows. Training using the accompaniment performances allows our model to learn some of the musicality these performances demonstrate. Since the soloist's interpretation must take precedence, we want to use this accompaniment interpretation only to the extent that it does not conflict with that of the soloist. We accomplish this by first beginning with the result of the accompaniment training described above. We use the practice room distributions , (the distributions on the {(Tn, O"n)} learned from the accompaniment data) , as the initial distributions , {J-L~ , ~~} . We then run the EM algorithm as described above now treating the currently available collection of solo performances as the observed data. During this phase, only those parameters relevant to the soloist's rhythmic interpretation will be modified significantly. Parameters describing the interpretation of a musical segment in which the soloist is mostly absent will be largely unaffected by the second training pass. Listen Update Composite Accomp Figure 3: At any given point in the performance we will have observed a collection of solo note times estimated estimated by Listen, and the accompaniment event times (the darkened circles). We compute the conditional distribution on the next unplayed accompaniment event, given these observations. This solo training actually happens over the course of a series of rehearsals. We first initialize our model to the practice room distribution by training with the accompaniment data. Then we iterate the process of creating a performance with our system, (described in the next section), extracting the sequence of solo note onset times in an off-line estimation process, and then retraining the model using all currently available solo performances. In our experience, only a few such rehearsals are necessary to train a system that responds gracefully and anticipates the soloist's rhythmic nuance where appropriate - generally less than 10. 5 Real Time Accompaniment The methodological key to our real-time accompaniment algorithm is the computation of (conditional) marginal distributions facilitated by the message-passing machinery of Bayesian networks. At any point during the performance some collection of solo notes and accompaniment notes will have been observed, as in Fig. 3. Conditioned on this information we can compute the distribution on the next unplayed accompaniment. The real-time computational requirement is limited by passing only the messages necessary to compute the marginal distribution on the pending accompaniment note. Once the conditional marginal distribution of the pending accompaniment note is calculated we schedule the note accordingly. Currently we schedule the note to be played at the conditional mean time, given all observed information, however other reasonable choices are possible. Note that this conditional distribution depends on all of the sources of information included in our model: The score information, all currently observed solo and accompaniment note times, and the rhythmic interpretations demonstrated by both the soloist and accompanist captured during the training phase. The initial scheduling of each accompaniment note takes place immediately after the previous accompaniment note is played. It is possible that a solo note will be detected before the pending accompaniment is played; in this event the pending accompaniment event is rescheduled by recomputing the its conditional distribution using the newly available information. The pending accompaniment note is rescheduled each time an additional solo note is detected until its currently scheduled time arrives, at which time it is finally played. In this way our accompaniment makes use of all currently available information. Does our system pass the musical equivalent of the Turing Test? We presume no more objectivity in answering this question than we would have in judging the merits of our other children. However, we believe that the level of musicality attained by our system is truly surprising, while the reliability is sufficient for live demonstration. We hope that the interested reader will form an independent opinion, even if different from ours, and to this end we have made musical examples demonstrating our progress available on the web page: http://fafner.math.umass.edu/musicplus_one. Acknowledgments This work supported by NSF grants IIS-998789 and IIS-0113496. References [1] Raphael C. (1999), "Automatic Segmentation of Acoustic Musical Signals Using Hidden Markov Models," IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 21, No.4, pp. 360-370. [2] Raphael C. (2001), "A Probabilistic Expert System for Automatic Musical Accompaniment," Journal of Computational and Graphical Statistics, vol. 10 no. 3, 487-512. [3] Raphael C. (2001), "Can the Computer Learn to Play Expressively?" Proceedings of Eighth International Workshop on Artificial Intelligence and Statistics, 113-120, Morgan Kauffman. [4] Raphael C. (2001), "Synthesizing Musical Accompaniments with Bayesian Belief Networks," Journal of New Music Research, vol. 30, no. 1, 59-67. [5] Spiegelhalter D., Dawid A. P., Lauritzen S., Cowell R. (1993), "Bayesian Analysis in Expert Systems," Statistical Science, Vol. 8, No.3, pp. 219-283. [6] Cowell R., Dawid A. P., Lauritzen S., Spiegelhalter D. (1999), "Probabilistic Networks and Expert Systems," Springer, New York. [7] Lauritzen S. L. (1995), "The EM Algorithm for Graphical Association Models with Missing Data," Computational Statistics and Data Analysis, Vol. 19, pp. 191-20l. [8] Lauritzen S. L. (1992), "Propagation of Probabilities, Means, and Variances in Mixed Graphical Association Models," Journal of the American Statistical Association, Vol. 87, No. 420, (Theory and Methods), pp. 1098-1108. [9] Lauritzen S. L. and F. Jensen (1999), "Stable Local Computation with Conditional Gaussian Distributions," Technical Report R-99-2014, Department of Mathematic Sciences, Aalborg University.
2035 |@word version:1 middle:1 compression:1 nd:1 retraining:1 gradual:1 bn:2 covariance:3 initial:5 series:2 uma:2 score:8 accompaniment:55 ours:1 past:4 surprising:1 must:4 ctn:1 treating:1 update:5 intelligence:2 slowing:1 accordingly:1 beginning:1 provides:1 math:2 location:1 constructed:1 become:1 manner:2 automatically:1 actual:2 provided:1 begin:6 estimating:1 lowest:1 backbone:1 developed:2 hindsight:1 grant:1 omit:1 positive:2 before:1 local:4 treat:1 challenging:1 limited:1 directed:1 acknowledgment:1 union:1 practice:3 significantly:1 composite:6 cannot:1 scheduling:1 context:1 live:5 equivalent:1 conventional:1 imposed:1 deterministic:1 demonstrated:1 musician:2 missing:1 baum:1 flexibly:1 duration:1 welch:1 immediately:1 notion:1 variation:4 feel:1 play:6 programming:1 dawid:2 located:1 asymmetric:1 labeled:4 bottom:2 role:2 observed:10 capture:1 removed:1 principled:2 respecting:1 dynamic:1 trained:2 segment:1 purely:1 efficiency:1 joint:2 various:4 train:1 fast:1 describe:4 effective:1 detected:4 artificial:1 tell:1 objectivity:1 quite:1 plausible:1 statistic:4 jointly:1 noisy:2 hoc:1 sequence:5 mg:1 raphael:6 neighboring:1 relevant:1 description:2 parent:1 asymmetry:1 requirement:1 produce:1 lauritzen:5 progress:1 indicate:1 concentrate:1 attribute:1 human:2 opinion:1 material:1 musically:2 expressively:2 precedence:1 stretch:1 around:2 considered:1 predict:2 mo:3 achieves:1 vary:1 estimation:2 currently:6 reflects:1 hope:1 clearly:1 gaussian:4 modified:1 rather:3 focus:1 methodological:1 rigid:1 mathematic:1 prescribing:1 hidden:5 interested:1 unobservable:3 overall:1 retaining:1 initialize:1 marginal:3 once:1 represents:2 unsupervised:1 constitutes:1 nearly:1 future:2 report:1 develops:1 duplicate:1 few:2 primarily:1 aalborg:1 composed:1 phase:4 consisting:1 message:5 workable:1 truly:1 arrives:1 solo:23 ambient:1 necessary:4 experience:1 machinery:1 owes:1 indexed:1 continuing:1 circle:3 recomputing:1 modeling:2 assignment:1 hundred:2 delay:1 successful:1 characterize:2 dependency:2 accomplish:1 anticipates:1 combined:1 confident:1 thoroughly:1 fundamental:1 amherst:2 international:1 probabilistic:4 off:2 synthesis:1 na:1 containing:1 creating:1 expert:3 american:1 sec:1 stabilize:1 explicitly:1 onset:6 depends:2 piece:6 performed:1 ad:1 complicated:1 accuracy:2 musical:25 variance:4 largely:1 wisdom:1 repeatability:1 bayesian:8 identification:1 produced:1 presume:1 unaffected:1 explain:1 rehearsal:5 pp:4 obvious:1 naturally:1 newly:1 treatment:1 massachusetts:1 recall:1 knowledge:8 listen:15 segmentation:1 schedule:2 actually:1 attained:1 follow:2 formulation:2 until:2 eqn:2 parse:1 christopher:1 web:1 propagation:1 accent:1 scheduled:1 perhaps:1 believe:1 true:2 evolution:1 hence:1 deal:1 during:5 ambiguous:1 rhythm:3 octave:1 stress:1 demonstrate:1 tn:22 performs:1 bring:1 common:1 quarter:1 conditioning:1 association:3 interpretation:15 relating:1 expressing:1 composition:1 refer:1 stretched:1 rd:1 automatic:2 grid:1 mathematics:1 similarly:1 reliability:1 stable:1 longer:1 etc:1 base:1 perspective:1 forcing:1 indispensable:1 occasionally:1 life:1 unplayed:2 captured:2 morgan:1 commentary:1 additional:1 performer:1 signal:8 ii:2 rv:2 sound:1 technical:1 adapt:1 equally:1 prediction:1 pitch:2 basic:2 denominator:1 essentially:1 addition:1 want:1 source:9 tend:1 call:1 extracting:1 iterate:1 isolation:1 listening:1 absent:2 passing:5 york:1 constitute:2 generally:1 latency:2 detailed:2 locally:1 documented:1 generate:1 http:1 nsf:1 judging:1 estimated:4 per:1 vol:6 express:1 diatonic:1 key:1 demonstrating:1 graph:6 run:1 facilitated:1 turing:1 respond:1 place:1 reasonable:1 reader:1 decision:1 layer:11 ct:2 played:7 occur:4 placement:1 constraint:2 precisely:2 generates:1 aspect:3 department:2 developing:1 describes:1 em:3 making:1 happens:1 explained:1 heart:1 equation:1 mutually:2 discus:1 describing:2 mind:1 merit:1 instrument:3 end:1 available:6 appropriate:1 tempo:11 responsive:1 voice:1 shortly:1 top:2 running:1 graphical:5 music:8 lmj:1 question:1 quantity:3 occurs:2 responds:2 darkened:3 hmm:3 gracefully:1 extent:1 nuance:1 length:3 modeled:2 index:1 demonstration:2 mostly:1 teach:1 negative:2 disparate:1 synthesizing:1 observation:3 markov:3 enabling:1 situation:1 precise:2 conflict:1 acoustic:7 engine:1 learned:4 established:1 address:1 able:1 pattern:1 kauffman:1 eighth:1 challenge:1 including:1 belief:1 rendition:2 event:17 synchronize:1 nth:1 representing:1 mn:3 improve:1 spiegelhalter:2 identifies:2 sn:3 speeding:1 relative:1 synchronization:2 mixed:1 reestimating:1 sufficient:1 consistent:1 playing:1 course:2 supported:1 jth:1 rhythmic:10 benefit:1 regard:1 boundary:2 calculated:1 collection:5 made:1 transaction:1 observable:5 midi:1 ml:1 assumed:1 learn:4 reasonably:1 pending:5 inherently:1 elastic:1 upstream:1 monophonic:1 arrow:1 noise:1 repeated:1 child:1 interpretive:1 fig:1 position:7 prominently:1 lie:2 answering:1 communicates:1 third:1 learns:1 down:1 specific:2 jensen:1 workshop:1 conditioned:1 demand:1 easier:1 suited:1 simply:1 contained:1 collectively:1 cowell:2 springer:1 corresponds:3 ma:1 conditional:13 viewed:1 sorted:1 room:4 rescheduled:2 change:7 included:1 specifically:1 microphone:1 f3n:3 pas:2 tendency:2 player:3
1,136
2,036
Minimax Probability Machine Gert R.G. Lanckriet* Department of EECS University of California, Berkeley Berkeley, CA 94720-1770 gert@eecs. berkeley.edu Laurent EI Ghaoui Department of EECS University of California, Berkeley Berkeley, CA 94720-1770 elghaoui@eecs.berkeley.edu Chiranjib Bhattacharyya Department of EECS University of California, Berkeley Berkeley, CA 94720-1776 chiru@eecs.berkeley.edu Michael I. Jordan Computer Science and Statistics University of California, Berkeley Berkeley, CA 94720-1776 jordan@cs.berkeley.edu Abstract When constructing a classifier, the probability of correct classification of future data points should be maximized. In the current paper this desideratum is translated in a very direct way into an optimization problem, which is solved using methods from convex optimization. We also show how to exploit Mercer kernels in this setting to obtain nonlinear decision boundaries. A worst-case bound on the probability of misclassification of future data is obtained explicitly. 1 Introduction Consider the problem of choosing a linear discriminant by minimizing the probabilities that data vectors fall on the wrong side of the boundary. One way to attempt to achieve this is via a generative approach in which one makes distributional assumptions about the class-conditional densities and thereby estimates and controls the relevant probabilities. The need to make distributional assumptions, however, casts doubt on the generality and validity of such an approach, and in discriminative solutions to classification problems it is common to attempt to dispense with class-conditional densities entirely. Rather than avoiding any reference to class-conditional densities, it might be useful to attempt to control misclassification probabilities in a worst-case setting; that is, under all possible choices of class-conditional densities. Such a minimax approach could be viewed as providing an alternative justification for discriminative approaches. In this paper we show how such a minimax programme can be carried out in the setting of binary classification. Our approach involves exploiting the following powerful theorem due to Isii [6], as extended in recent work by Bertsimas ? http://robotics.eecs.berkeley.edur gert/ and Sethuraman [2]: where y is a random vector, where a and b are constants, and where the supremum is taken over all distributions having mean y and covariance matrix ~y. This theorem provides us with the ability to bound the probability of misclassifying a point, without making Gaussian or other specific distributional assumptions. We will show how to exploit this ability in the design of linear classifiers. One of the appealing features of this formulation is that one obtains an explicit upper bound on the probability of misclassification of future data: 1/(1 + rP). A second appealing feature of this approach is that, as in linear discriminant analysis [7], it is possible to generalize the basic methodology, utilizing Mercer kernels and thereby forming nonlinear decision boundaries. We show how to do this in Section 3. The paper is organized as follows: in Section 2 we present the minimax formulation for linear classifiers, while in Section 3 we deal with kernelizing the method. We present empirical results in Section 4. 2 Maximum probabilistic decision hyperplane In this section we present our minimax formulation for linear decision boundaries. Let x and y denote random vectors in a binary classification problem, with mean vectors and covariance matrices given by x '" (x, ~x) and y '" (y, ~y) , respectively, where ""," means that the random variable has the specified mean and covariance matrix but that the distribution is otherwise unconstrained. Note that x, x , y , Y E JRn and ~x, ~y E JRnxn. We want to determine the hyperplane aT z = b (a, z E JRn and b E JR) that separates the two classes of points with maximal probability with respect to all distributions having these means and covariance matrices. This boils down to: max a s.t. a ,a ,b inf Pr{ aT x 2: b} 2: a (2) or, max a a,a,b s.t. 1 - a 2: sup Pr{ aT x 1- a :s b} (3) 2: sup Pr{aT y 2: b} . Consider the second constraint in (3). Recall the result of Bertsimas and Sethuraman [2]: 1 supPr{aTY2:b}=-d2' with 1+ d2 = inf (Y_Yf~y-1(y_y) (4) aTy?b We can write this as d2 = infcTw>d w Tw, where w = ~y -1 /2 (y_y), c T = aT~y 1/2 and d = b - aTy. To solve this,-first notice that we can assume that aTy :S b (i.e. y is classified correctly by the decision hyperplane aT z = b): indeed, otherwise we would find d2 = 0 and thus a = 0 for that particular a and b, which can never be an optimal value. So, d> o. We then form the Lagrangian: ?(w, >.) = w T w + >.(d - c T w), (5) which is to be maximized with respect to A 2: 0 and minimized with respect to w . At the optimum, 2w = AC and d = c T W , so A = and w = This yields: -!#c c%c. (6) Using (4), the second constraint in (3) becomes 1-0: 2: 1/(I+d2 ) or ~ 2: 0:/(1-0:). Taking (6) into account, this boils down to: b-aTY2:,,(o:)/aT~ya V where ,,(0:)=) 0: 1-0: (7) We can handle the first constraint in (3) in a similar way (just write aT x ::::: b as _aT x 2: -b and apply the result (7) for the second constraint). The optimization problem (3) then becomes: max 0: -b + aTx 2: ,,(o:)JaT~xa s.t. a ,a,b b - aTy 2: (8) "(o:h/aT~ya. Because "(0:) is a monotone increasing function of 0:, we can write this as: max" (9) s.t. ""a,b b - aTy 2: "JaT~ya. From both constraints in (9), we get aTy + "JaT~ya::::: b::::: aTx - "JaT~xa, (10) which allows us to eliminate b from (9): max" I<,a s.t. aTy + "JaT~ya::::: aTx - "JaT~xa. (11) Because we want to maximize ", it is obvious that the inequalities in (10) will become equalities at the optimum. The optimal value of b will thus be given by (12) where a* and "* are the optimal values of a and " respectively. Rearranging the constraint in (11), we get: aT(x - y) 2:" (JaT~xa+ JaT~ya). (13) The above is positively homogeneous in a: if a satisfies (13), sa with s E 114 also does. Furthermore, (13) implies aT(x - y) 2: O. Thus, we can restrict a to be such that aT(x - y) = 1. The optimization problem (11) then becomes max" I<,a s.t. ~ 2: JaT~xa + JaT~ya (14) a T (x-Y)=I , which allows us to eliminate ,,: m~n JaT~xa + JaT~ya s.t. aT(x - y) = 1, (15) or, equivalently (16) This is a convex optimization problem, more precisely a second order cone program (SOCP) [8,5]. Furthermore, notice that we can write a = ao +Fu, where U E Il~n-l, ao = (x - y)/llx - y112, and F E IRnx (n-l) is an orthogonal matrix whose columns span the subspace of vectors orthogonal to x - y. Using this we can write (16) as an unconstrained SOCP: (17) We can solve this problem in various ways, for example using interior-point methods for SOCP [8], which yield a worst-case complexity of O(n 3 ). Of course, the first and second moments of x, y must be estimated from data, using for example plug-in estimates X, y, :Ex, :E y for respectively x, y, ~x, ~y. This brings the total complexity to O(ln 3 ), where l is the number of data points. This is the same complexity as the quadratic programs one has to solve in support vector machines. In our implementations, we took an iterative least-squares approach, which is based on the following form , equivalent to (17): (18) At iteration k , we first minimize with respect to 15 and E by setting 15k = II~x 1/2(ao + Ek = II~y 1/2(ao + Fu k - 1)112. Then we minimize with respect to U by solving a least squares problem in u for 15 = 15k and E = Ek, which gives us Uk. Because in both update steps the objective of this COP will not increase, the iteration will converge to the global minimum II~xl/2(ao + Fu*)112 + II~yl /2(ao + Fu*)lb with u* an optimal value of u. Fu k - d112 and We then obtain a* as ao + Fu* and b* from (12) with "'* = l/h/ar~xa* + Jar~ya*). Classification of a new data point Zn ew is done by evaluating sign( a;; Zn ew - b*): if this is +1, Zn ew is classified as from class x, otherwise Zn ew is classified as from class y. It is interesting to see what happens if we make distributional assumptions; in particular, let us assume that x "" N(x, ~x) and y "" N(y, ~y). This leads to the following optimization problem: max a o:, a ,b S.t. -b + aTx ::::: <I>-l(a)JaT~xa (19) where <I>(z) is the cumulative distribution function for a standard normal Gaussian distribution. This has the same form as (8), but now with ",(a) = <I>-l(a) instead of ",(a) = Vl~a (d. a result by Chernoff [4]). We thus solve the same optimization problem (a disappears from the optimization problem because ",(a) is monotone increasing) and find the same decision hyperplane aT z = b. The difference lies in the value of a associated with "'*: a will be higher in this case, so the hyperplane will have a higher predicted probability of classifying future data correctly. Kernelization 3 In this section we describe the "kernelization" of the minimax approach described in the previous section. We seek to map the problem to a higher dimensional feature space ]Rf via a mapping cP : ]Rn 1-+ ]Rf, such that a linear discriminant in the feature space corresponds to a nonlinear discriminant in the original space. To carry out this programme, we need to try to reformulate the minimax problem in terms of a kernel function K(Z1' Z2) = cp(Z1)T CP(Z2) satisfying Mercer's condition. 1-+ cp(x) ""' (cp(X) , ~cp(x)) and Y 1-+ cp(y) ""' (cp(y) , ~cp(y)) where {Xi}~1 and {Yi}~1 are training data points in the classes Let the data be mapped as x corresponding to x and Y respectively. The decision hyperplane in ]Rf is then given by aT cp(Z) = b with a, cp(z) E ]Rf and b E ]R. In ]Rf, we need to solve the following optimization problem: mln Jr-aT-~-cp-(-x)-a + JaT~cp(y)a aT (cp(X) - cp(y)) = 1, s.t. (20) where, as in (12), the optimal value of b will be given by b* = a; cp(x) - "'*Jar~cp(x)a* + "'*Jar~cp(y)a*, = a; cp(y) (21) where a* and "'* are the optimal values of a and '" respectively. However, we do not wish to solve the COP in this form, because we want to avoid using f or cp explicitly. If a has a component in ]Rf which is orthogonal to the subspace spanned by CP(Xi), i = 1,2, ... , N x and CP(Yi), i = 1,2, ... , Ny, then that component won't affect the objective or the constraint in (20) . This implies that we can write a as N. Ny (22) a = LaiCP(Xi) + L;)jCP(Yj). i=1 j=1 Substituting expression (22) for a and estimates = 2:~1 CP(Xi) , ;p(Y) = 1 Ny 1 N. .....--.. .....--.. T Ny 2:i=l cp(Yi), ~cp(x) - N. 2: i=1 (cp(Xi) - cp(X)) (cp(Xi) - cp(x)) and ~cp(y) - ;Pw A J y N J. _ A .....--.. _ .....--.. 2:i~1(CP(Yi) - cp(y))(cp(Yi) - cp(y))T for the means and the covariance matri- ces in the objective and the constraint of the optimization problem (20), we see that both the objective and the constraints can be written in terms of the kernel function K(Zl' Z2) = CP(Z1)T cp(Z2) . We obtain: T - - (23) "f (k x - ky) = 1, T J. - N N . - [a1 a2 ... aN. ;)1 ;)2 ... ;)Ny l , kx E ]R .+ y WIth [kxl i = 2:f;1 K(xj, Zi), ky E ]RN. +Ny with [kyl i = y 2:f~l K(Yj, Zi), Zi = Xi for where "f = J i = 1,2, ... ,Nx and Zi = Yi - N. for i = N x as: + 1, N x + 2, ... ,Nx + Ny . K is defined (Kx -IN.~~) = (*x) (24) Ky -lNy ky Ky where 1m is a column vector with ones of dimension m. Kx and Ky contain respectively the first N x rows and the last Ny rows of the Gram matrix K (defined as Kij = cp(zdTcp(zj) = K(Zi,Zj)). We can also write (23) as K= - Kx I m~n II ~"f12 Ky I + II.jlV;"f 12 s.t. T - - "f (kx - ky) = 1, (25) which is a second order cone program (SOCP) [5] that has the same form as the SOCP in (16) and can thus be solved in a similar way. Notice that, in this case, the optimizing variable is "f E ~Nz +Ny instead of a E ~n. Thus the dimension of the optimization problem increases, but the solution is more powerful because the kernelization corresponds to a more complex decision boundary in ~n . Similarly, the optimal value b* of b in (21) will then become (26) "'* are the optimal values of "f and", respectively. Once "f* is known, we get "'* = 1/ ( J~z "f;K~Kx"f* + J~y "f;K~Ky"f* ) and then where "f* and b* from (26). Classification of a new data point Znew is then done by evaluating sign(a; <p(znew) -b*) = sign ( (L~l+Ny b*]iK(Zi, Znew) ) - b*) (again only in terms of the kernel function): if this is + 1, Znew is classified as from class x , otherwise Znew is classified as from class y. 4 Experiments In this section we report the results of experiments that we carried out to test our algorithmic approach. The validity of 1 - a as the worst case bound on the probability of misclassification of future data is checked, and we also assess the usefulness of the kernel trick in this setting. We compare linear kernels and Gaussian kernels. Experimental results on standard benchmark problems are summarized in Table 1. The Wisconsin breast cancer dataset contained 16 missing examples which were not used. The breast cancer, pima, diabetes, ionosphere and sonar data were obtained from the VCI repository. Data for the twonorm problem data were generated as specified in [3]. Each dataset was randomly partitioned into 90% training and 10% test sets. The kernel parameter (u) for the Gaussian kernel (e-llx-yI12/,,) was tuned using cross-validation over 20 random partitions. The reported results are the averages over 50 random partitions for both the linear kernel and the Gaussian kernel with u chosen as above. The results are comparable with those in the existing literature [3] and with those obtained with Support Vector Machines. Also, we notice that a is indeed smaller Table 1: a and test-set accuracy (TSA) compared to BPB (best performance in [3]) and to the performance of an SVM with linear kernel (SVML) and an SVM with Gaussian kernel (SVMG) Dataset Twonorm Breast cancer Ionosphere Pima diabetes Sonar Linear a 80.2 % 84.4 % 63.3 % 31.2 % 62.4 % kernel TSA: 96.0 % 97.2 % 85.4 % 73.8 % 75.1 % BPB SVML SVMG Gaussian kernel TSA: a 83.6 % 97.2 % 96.3 % 95.6 % 97.4 % 92.7 % 97.3 % 96.8 % 92.6 % 98.5 % 89.9 % 93.0 % 93.7 % 87.8 % 91.5 % 33.0 % 74.6 % 76.1 % 70.1 % 75.3 % 87.1 % 89.8 % 75.9 % 86.7 % than the test-set accuracy in all cases. Furthermore, a is smaller for a linear decision boundary then for the nonlinear decision boundary obtained via the Gaussian kernel. This clearly shows that kernelizing the method leads to more powerful decision boundaries. 5 Conclusions The problem of linear discrimination has a long and distinguished history. Many results on misclassification rates have been obtained by making distributional assumptions (e.g., Anderson and Bahadur [1]) . Our results , on the other hand, make use of recent work on moment problems and semidefinite optimization to obtain distribution-free results for linear discriminants. We have also shown how to exploit Mercer kernels to generalize our algorithm to nonlinear classification. The computational complexity of our method is comparable to the quadratic program that one has to solve for the support vector machine (SVM). While we have used a simple iterative least-squares approach, we believe that there is much to gain from exploiting analogies to the SVM and developing specialized, more efficient optimization procedures for our algorithm, in particular tools that break the data into subsets. The extension towards large scale applications is a current focus of our research, as is the problem of developing a variant of our algorithm for multiway classification and function regression . Also the statistical consequences of using plug-in estimates for the mean vectors and covariance matrices needs to be investigated. Acknowledgements We would like to acknowledge support from ONR MURI N00014-00-1-0637, from NSF grants IIS-9988642 and ECS-9983874 and from the Belgian American Educational Foundation. References [1] Anderson, T . W . and Bahadur, R. R . (1962) Classification into two multivariate Normal distributions with different covariance matrices. Annals of Mathematical Statistics 33(2): 420-431. [2] Bertsimas, D. and Sethuraman, J. (2000) Moment problems and semidefinite optimization. Handbook of Semidefinite Optimization 469-509, Kluwer Academic Publishers. [3] Breiman L. (1996) Arcing classifiers. Technical Report 460 , Statistics Department, University of California, December 1997. [4] Chernoff H. (1972) The selection of effective attributes for deciding between hypothesis using linear discriminant functions. In Frontiers of Pattern Recognition, (S. Watanabe, ed.), 55-60. New York: Academic Press. [5] Boyd, S. and Vandenberghe, L. (2001) Convex Optimization. Course notes for EE364, Stanford University. Available at http://www . stanford. edu/ class/ee364. [6] Isii, K. (1963) On the sharpness of Chebyshev-type inequalities. Math . 14: 185-197. Ann. Inst. Stat. [7] Mika, M. Ratsch, G., Weston, J., SchOikopf, B., and Mii11er, K.-R. (1999) Fisher discriminant analysis with kernels. In Neural Networks for Signal Processing IX, 41- 48 , New York: IEEE Press. [8] Nesterov , Y. and Nemirovsky, A. (1994) Interior Point Polynomial Methods in Convex Programming: Theory and Applications. Philadelphia, PA: SIAM.
2036 |@word repository:1 pw:1 polynomial:1 d2:5 seek:1 nemirovsky:1 covariance:7 thereby:2 carry:1 moment:3 tuned:1 bhattacharyya:1 existing:1 current:2 z2:4 must:1 written:1 partition:2 update:1 tsa:3 discrimination:1 generative:1 mln:1 provides:1 math:1 mathematical:1 direct:1 become:2 ik:1 indeed:2 increasing:2 becomes:3 what:1 berkeley:13 classifier:4 wrong:1 uk:1 control:2 zl:1 grant:1 consequence:1 laurent:1 might:1 mika:1 nz:1 yj:2 atx:4 procedure:1 empirical:1 boyd:1 get:3 interior:2 selection:1 www:1 equivalent:1 map:1 lagrangian:1 missing:1 educational:1 convex:4 sharpness:1 utilizing:1 spanned:1 vandenberghe:1 handle:1 gert:3 justification:1 annals:1 programming:1 homogeneous:1 hypothesis:1 lanckriet:1 trick:1 diabetes:2 pa:1 satisfying:1 recognition:1 distributional:5 muri:1 solved:2 worst:4 complexity:4 dispense:1 nesterov:1 solving:1 translated:1 various:1 describe:1 effective:1 choosing:1 whose:1 stanford:2 solve:7 otherwise:4 ability:2 statistic:3 cop:2 took:1 maximal:1 relevant:1 achieve:1 ky:9 exploiting:2 optimum:2 ac:1 stat:1 sa:1 c:1 involves:1 implies:2 predicted:1 correct:1 attribute:1 ao:7 extension:1 frontier:1 normal:2 deciding:1 mapping:1 algorithmic:1 substituting:1 a2:1 tool:1 clearly:1 gaussian:8 rather:1 avoid:1 breiman:1 arcing:1 focus:1 inst:1 vl:1 eliminate:2 classification:9 once:1 never:1 having:2 chernoff:2 future:5 minimized:1 report:2 randomly:1 attempt:3 semidefinite:3 fu:6 belgian:1 orthogonal:3 kij:1 column:2 ar:1 zn:4 subset:1 usefulness:1 reported:1 eec:7 density:4 siam:1 twonorm:2 probabilistic:1 yl:1 michael:1 again:1 ek:2 american:1 doubt:1 account:1 socp:5 summarized:1 explicitly:2 try:1 break:1 sup:2 ass:1 il:1 square:3 minimize:2 f12:1 accuracy:2 maximized:2 yield:2 generalize:2 classified:5 history:1 checked:1 ed:1 obvious:1 associated:1 boil:2 gain:1 dataset:3 recall:1 organized:1 higher:3 methodology:1 formulation:3 done:2 generality:1 furthermore:3 just:1 xa:8 anderson:2 hand:1 ei:1 vci:1 nonlinear:5 brings:1 believe:1 validity:2 contain:1 equality:1 deal:1 won:1 cp:37 common:1 specialized:1 discriminants:1 kluwer:1 llx:2 unconstrained:2 similarly:1 multiway:1 multivariate:1 recent:2 optimizing:1 inf:2 n00014:1 inequality:2 binary:2 onr:1 yi:6 kyl:1 aty2:2 minimum:1 determine:1 maximize:1 converge:1 signal:1 ii:7 technical:1 academic:2 plug:2 cross:1 long:1 a1:1 desideratum:1 basic:1 variant:1 breast:3 regression:1 aty:7 iteration:2 kernel:19 robotics:1 want:3 bpb:2 ratsch:1 publisher:1 december:1 jordan:2 affect:1 xj:1 zi:6 restrict:1 y112:1 chebyshev:1 expression:1 schoikopf:1 york:2 useful:1 jrn:2 http:2 bahadur:2 misclassifying:1 zj:2 notice:4 nsf:1 sign:3 svml:2 estimated:1 correctly:2 write:7 ce:1 bertsimas:3 monotone:2 cone:2 znew:5 powerful:3 decision:11 comparable:2 entirely:1 bound:4 matri:1 quadratic:2 constraint:9 precisely:1 span:1 department:4 developing:2 jr:2 smaller:2 partitioned:1 appealing:2 tw:1 making:2 happens:1 pr:3 ghaoui:1 taken:1 ln:1 chiranjib:1 available:1 apply:1 distinguished:1 kernelizing:2 alternative:1 isii:2 rp:1 original:1 exploit:3 objective:4 subspace:2 separate:1 mapped:1 nx:2 discriminant:6 reformulate:1 providing:1 minimizing:1 equivalently:1 pima:2 jat:14 design:1 implementation:1 upper:1 benchmark:1 acknowledge:1 extended:1 rn:2 lb:1 cast:1 specified:2 z1:3 california:5 pattern:1 kxl:1 elghaoui:1 program:4 yi12:1 rf:6 max:7 lny:1 misclassification:5 minimax:7 disappears:1 sethuraman:3 carried:2 philadelphia:1 literature:1 acknowledgement:1 wisconsin:1 interesting:1 analogy:1 validation:1 foundation:1 mercer:4 classifying:1 row:2 cancer:3 course:2 last:1 free:1 side:1 fall:1 taking:1 boundary:8 dimension:2 evaluating:2 cumulative:1 gram:1 programme:2 ec:1 obtains:1 supremum:1 global:1 handbook:1 discriminative:2 xi:7 jrnxn:1 iterative:2 sonar:2 table:2 ca:4 rearranging:1 investigated:1 complex:1 constructing:1 positively:1 ny:10 watanabe:1 explicit:1 wish:1 xl:1 lie:1 ix:1 theorem:2 down:2 specific:1 svm:4 ionosphere:2 kx:6 jar:3 forming:1 contained:1 corresponds:2 satisfies:1 weston:1 chiru:1 conditional:4 viewed:1 ann:1 towards:1 jlv:1 fisher:1 hyperplane:6 total:1 experimental:1 ya:9 ew:4 support:4 kernelization:3 avoiding:1 ex:1
1,137
2,037
Escaping the Convex Hull with Extrapolated Vector Machines. Patrick Haffner AT&T Labs-Research, 200 Laurel Ave, Middletown, NJ 07748 haffner@research.att.com Abstract Maximum margin classifiers such as Support Vector Machines (SVMs) critically depends upon the convex hulls of the training samples of each class, as they implicitly search for the minimum distance between the convex hulls. We propose Extrapolated Vector Machines (XVMs) which rely on extrapolations outside these convex hulls. XVMs improve SVM generalization very significantly on the MNIST [7] OCR data. They share similarities with the Fisher discriminant: maximize the inter-class margin while minimizing the intra-class disparity. 1 Introduction Both intuition and theory [9] seem to support that the best linear separation between two classes is the one that maximizes the margin. But is this always true? In the example shown in Fig.(l), the maximum margin hyperplane is Wo; however , most observers would say that the separating hyperplane WI has better chances to generalize, as it takes into account the expected location of additional training sam- ? ? ? ? ? ? ? ? ? ? ? f\J:- ? ? ? ? ? ? ? ? ? ? ?? . . , --.Q. ~"- _ ~ , ................ ~x~... --------------- .. ?? .' W 1 ~--------------- ? K~ ???????????0 '''-,,- / 0- 0 00 00 0 0'\ "OW-0_ o_ _o- o- o::-o . .. . ....... ............ . (} Figure 1: Example of separation where the large margin is undesirable. The convex hull and the separation that corresponds to the standard SVM use plain lines while the extrapolated convex hulls and XVMs use dotted lines. pIes. Traditionally, to take this into account, one would estimate the distribution of the data. In this paper, we just use a very elementary form of extrapolation ("the poor man variance") and show that it can be implemented into a new extension to SVMs that we call Extrapolated Vector Machines (XVMs). 2 Adding Extrapolation to Maximum Margin Constraints This section states extrapolation as a constrained optimization problem and computes a simpler dual form. Take two classes C+ and C_ with Y+ = +1 and Y_ = -1 1 as respective targets. The N training samples {(Xi, Yi); 1 ::::; i ::::; N} are separated with a margin p if there exists a set of weights W such that Ilwll = 1 and Vk E {+, -}, Vi E Ck, Yk(w,xi+b) 2: p (1) SVMs offer techniques to find the weights W which maximize the margin p. Now, instead of imposing the margin constraint on each training point, suppose that for two points in the same class Ck, we require any possible extrapolation within a range factor 17k 2: 0 to be larger than the margin: Vi,j E Ck , V)" E [-17k, l+17k], Yk (W.()"Xi + (l-)")Xj) + b) 2: P (2) It is sufficient to enforce the constraints at the end of the extrapolation segments, and (3) Keeping the constraint over each pair of points would result in N 2 Lagrange multipliers. But we can reduce it to a double constraint applied to each single point. If follows from Eq.(3) that: (4) (5) We consider J.Lk = max (Yk(W.Xj)) and Vk = min (Yk(W.Xj)) as optimization varilEC. lEC. abIes. By adding Eq.(4) and (5), the margin becomes 2p = L ((17k+ 1)vk - 17kJ.Lk) = k L (Vk -17dJ.Lk - Vk)) (6) k Our problem is to maximize the margin under the double constraint: Vi E Ck , Vk ::::; Yk(W.Xi) ::::; J.Lk In other words, the extrapolated margin maximization is equivalent to squeezing the points belonging to a given class between two hyperplanes. Eq.(6) shows that p is maximized when Vk is maximized while J.Lk - Vk is minimized. Maximizing the margin over J.Lk , Vk and following dual problem: W with Lagrangian techniques gives us the (7) lIn this paper, it is necessary to index the outputs y with the class k rather than the more traditional sample index i, as extrapolation constraints require two examples to belong to the same class. The resulting equations are more concise, but harder to read. Compared to the standard SVM formulation, we have two sets of support vectors. Moreover, the Lagrange multipliers that we chose are normalized differently from the traditional SVM multipliers (note that this is one possible choice of notation, see Section.6 for an alternative choice). They sum to 1 and allow and interesting geometric interpretation developed in the next section. 3 Geometric Interpretation and Iterative Algorithm For each class k, we define the nearest point to the other class convex hull along the direction of w: Nk = I:iECk f3iXi. Nk is a combination of the internal support vectors that belong to class k with f3i > O. At the minimum of (7), because they correspond to non zero Lagrange multipliers, they fallon the internal margin Yk(W,Xi) = Vk; therefore, we obtain Vk = Ykw.Nk? Similarly, we define the furthest point Fk = I: i ECk ~i Xi' Fk is a combination of the external support vectors, and we have flk = Ykw.Fk. The dual problem is equivalent to the distance minimization problem IILYk ((1Jk+I)Nk _1Jk F k)11 k min Nk ,Fk EHk where 1{k 2 is the convex hull containing the examples of class k. It is possible to solve this optimization problem using an iterative Extrapolated Convex Hull Distance Minimization (XCHDM) algorithm. It is an extension of the Nearest Point [5] or Maximal Margin Percept ron [6] algorithms. An interesting geometric interpretation is also offered in [3]. All the aforementioned algorithms search for the points in the convex hulls of each class that are the nearest to each other (Nt and No on Fig.I) , the maximal margin weight vector w = Nt - N o-' XCHDM look for nearest points in the extrapolated convex hulls (X+ I and X-I on Fig.I). The extrapolated nearest points are X k = 1JkNk - 1JkFk' Note that they can be outside the convex hull because we allow negative contribution from external support vectors. Here again, the weight vector can be expressed as a difference between two points w = X+ - X - . When the data is non-separable, the solution is trivial with w = O. With the double set of Lagrange multipliers, the description of the XCHDM algorithm is beyond the scope of this paper. XCHDM with 1Jk = 0 are simple SVMs trained by the same algorithm as in [6]. An interesting way to follow the convergence of the XCHDM algorithm is the following. Define the extrapolated primal margin 1'; = 2p = L ((1Jk+ I )vk - 1Jkflk) k and the dual margin 1'; = IIX+ - X-II Convergence consists in reducing the duality gap 1'~ -1'; down to zero. In the rest of the paper, we will measure convergence with the duality ratio r = 1'~ . 1'2 To determine the threshold to compute the classifier output class sign(w.x+b) leaves us with two choices. We can require the separation to happen at the center of the primal margin, with the primal threshold (subtract Eq.(5) from Eq.(4)) bl = 1 -2" LYk ((1Jk+ I )vk-1JkJ.lk) k or at the center of the dual margin, with the dual threshold b2 = - ~w. 2:)(T}k+1)Nk - T}kFk) = - ~ (IIx+ 112 -lix-in k Again, at the minimum, it is easy to verify that b1 = b2 . When we did not let the XCHDM algorithm converge to the minimum, we found that b1 gave better generalization results. Our standard stopping heuristic is numerical: stop when the duality ratio gets over a fixed value (typically between 0.5 and 0.9). The only other stopping heuristic we have tried so far is based on the following idea. Define the set of extrapolated pairs as {(T}k+1)Xi -T}kXj; 1 :S i,j :S N}. Convergence means that we find extrapolated support pairs that contain every extrapolated pair on the correct side of the margin. We can relax this constraint and stop when the extrapolated support pairs contain every vector. This means that 12 must be lower than the primal true margin along w (measured on the non-extrapolated data) 11 = y+ + Y - . This causes the XCHDM algorithm to stop long before 12 reaches Ii and is called the hybrid stopping heuristic. 4 Beyond SVMs and discriminant approaches. Kernel Machines consist of any classifier of the type f(x) = L:i Yi(Xi K(x, Xi). SVMs offer one solution among many others, with the constraint (Xi > O. XVMs look for solutions that no longer bear this constraint. While the algorithm described in Section 2 converges toward a solution where vectors act as support of margins (internal and external), experiments show that the performance of XVMs can be significantly improved if we stopped before full convergence. In this case, the vectors with (Xi =/: 0 do not line up onto any type of margin, and should not be called support vectors. The extrapolated margin contains terms which are caused by the extrapolation and are proportional to the width of each class along the direction of w. We would observe the same phenomenon if we had trained the classifier using Maximum Likelihood Estimation (MLE) (replace class width with variance). In both MLE and XVMs, examples which are the furthest from the decision surface play an important role. XVMs suggest an explanation why. Note also that like the Fisher discriminant , XVMs look for the projection that maximizes the inter-class variance while minimizing the intra-class variances. 5 Experiments on MNIST The MNIST OCR database contains 60,000 handwritten digits for training and 10,000 for testing (the testing data can be extended to 60,000 but we prefer to keep unseen test data for final testing and comparisons). This database has been extensively studied on a large variety of learning approaches [7]. It lead to the first SVM "success story" [2], and results have been improved since then by using knowledge about the invariance of the data [4]. The input vector is a list of 28x28 pixels ranging from 0 to 255. Before computing the kernels , the input vectors are normalized to 1: x = II~II' Good polynomial kernels are easy to define as Kp(x, y) = (x.y)P. We found these normalized kernels to outperform the unnormalized kernels Kp(x, y) = (a(x.y)+b)P that have been traditionally used for the MNIST data significantly. For instance, the baseline error rate with K4 is below 1.2%, whereas it hovers around 1.5% for K4 (after choosing optimal values for a and b)2. We also define normalized Gaussian kernels: Kp(x, y) = exp ( - ~ Ilx - y112) = [exp (x.y- 1)JP. (8) Eq.(8) shows how they relate to normalized polynomial kernels: when x.y ? 1, Kp and Kp have the same asymptotic behavior. We observed that on MNIST, the performance with Kp is very similar to what is obtained with unnormalized Gaussian kernels Ku(x , y) = exp _(X~Y)2. However, they are easier to analyze and compare to polynomial kernels. MNIST contains 1 class per digit, so the total number of classes is M=10. To combine binary classifiers to perform multiclass classifications, the two most common approaches were considered . ? In the one-vs-others case (lvsR) , we have one classifier per class c, with the positive examples taken from class c and negative examples form the other classes. Class c is recognized when the corresponding classifier yields the largest output . ? In the one-vs-one case (lvs1), each classifier only discriminates one class from another: we need a total of (MU:;-l) = 45 classifiers. Despite the effort we spent on optimizing the recombination of the classifiers [8] 1vsR SVMs (Table 1) perform significantly better than 1vs1 SVMs (Table 2). 4 3, For each trial, the number of errors over the 10,000 test samples (#err) and the total number of support vectors( #SV) are reported. As we only count SVs which are shared by different classes once, this predicts the test time. For instance, 12,000 support vectors mean that 20% of the 60,000 vectors are used as support. Preliminary experiments to choose the value of rJk with the hybrid criterion show that the results for rJk = 1 are better than rJk = 1.5 in a statistically significant way, and slightly better than rJk = 0.5. We did not consider configurations where rJ+ f; rJ -; however, this would make sense for the assymetrical 1vsR classifiers. XVM gain in performance over SVMs for a given configuration ranges from 15% (1 vsR in Table 3) to 25% (1 vs1 in Table 2). 2This may partly explain a nagging mystery among researchers working on MNIST: how did Cortes and Vapnik [2] obtain 1.1% error with a degree 4 polynomial ? 3We compared the Max Wins voting algorithm with the DAGSVM decision tree algorithm and found them to perform equally, and worse than 1vsR SVMs. This is is surprising in the light of results published on other tasks [8] , and would require further investigations beyond the scope of this paper. 4Slightly better performance was obtained with a new algorithm that uses the incremental properties of our training procedure (this is be the performance reported in the tables). In a transductive inference framework , treat the test example as a training example: for each of the M possible labels, retrain the M among (M(":-l) classifiers that use examples with such label. The best label will be the one that causes the smallest increase in the multiclass margin p such that it combines the classifier margins pc in the following manner ~= ,,~ 2 ~ 2 P c~M Pc The fact that this margin predicts generalization is "justified" by Theorem 1 in [8]. Kernel K3 K4 K5 Kg [(2 [(4 K5 0.40 #err #SV 136 8367 127 8331 125 8834 136 13002 147 9014 125 8668 125 8944 Duality Ratio stop 0.75 #err #SV 136 11132 117 11807 119 12786 137 18784 128 11663 12222 119 125 12852 # err 132 119 119 141 131 117 125 0.99 #SV 13762 15746 17868 25953 13918 16604 18085 Table 1: SVMs on MNIST with 10 1vsR classifiers Kernel K3 K4 K5 SVM/ratio at 0.99 # err #SV 138 11952 135 13526 191 13526 XVM/Hybrid # err #SV 117 17020 110 16066 114 15775 Table 2: SVMjXVM on MNIST with 45 1vs1 classifiers The 103 errors obtained with K4 and r = 0.5 in Table 3 represent only about 1% error: t his is t he lowest error ever reported for any learning technique without a priori knowledge about the fact that t he input data corresponds to a pixel map (the lowest reproducible error previously reported was 1.2% with SVMs and polynomials of degree 9 [4], it could be reduced to 0.6% by using invariance properties of the pixel map). The downside is that XVMs require 4 times as many support vectors as standards SVMs. Table 3 compares stopping according to t he duality ratio and t he hybrid criterion. With t he duality ratio, the best performance is most often reached with r = 0.50 (if t his happens to be consistent ly true, validation data to decide when to stop could be spared). The hybrid criterion does not require validation data and yields errors that, while higher than the best XVM, are lower than SVMs and only require a few more support vectors. It takes fewer iterations to train than SVMs. One way to interpret this hybrid stopping criterion is that we stop when interpolation in some (but not all) directions account for all non-interpolated vectors. This suggest t hat interpolation is only desirable in a few directions. XVM gain is stronger in the 1vs 1 case (Table 2). This suggests that extrapolating on a convex hull that contains several different classes (in the 1vsR case) may be undesirable. Kernel K3 K4 K5 Kg K2 [(4 0.40 # err #SV 118 46662 112 40274 109 36912 128 35809 114 43909 108 36980 Duality Ratio stop 0. 50 # err #SV 111 43819 43132 103 44226 106 126 39462 114 46905 111 40329 0.75 # err #SV 116 50216 110 52861 110 49383 131 50233 114 53676 114 51088 Hybrid. Stop Crit. # err #SV 125 20604 18002 107 17322 107 125 19218 119 20152 16895 108 Table 3: XVMs on MNIST wit h 10 1vsR classifiers 6 The Soft Margin Case MNIST is characterized by the quasi-absence of outliers, so to assume that the data is fully separable does not impair performance at all. To extend XVMs to non-separable data, we first considered the traditional approaches of adding slack variables to allow margin constraints to be violated. The most commonly used approach with SVMs adds linear slack variables to the unitary margin. Its application to the XVM requires to give up the weight normalization constraint, so that the usual unitary margin can be used in the constraints [9] . Compared to standard SVMs, a new issue to tackle is the fact that each constraint corresponds to a pair of vectors: ideally, we should handle N 2 slack variables ~ij. To have linear constraints that can be solved with KKT, we need to have the decomposition ~ij = ('T}k+1)~i+'T}k~; (factors ('T}k+1) and 'T}k are added here to ease later simplifications). Similarly to Eq.(3), the constraint on the extrapolation from any pair of points is Vi,j E Ck, Yk (w. (('T}k+1)xi - 'T}kXj) +b) 2: 1 - ('T}k+1)~i - 'T}k~; with ~i'~; 2: 0 Introducing J.tk = max (Yk(w,xj+b) - ~;) and Vk = min (Yk(W,Xi+b) .ECk JECk + ~i)' (9) we ob- tain the simpler double constraint Vi E Ck , Vk -~i ~ Yk(W,Xi+b) ~ J.tk+~; with ~i'~; 2: 0 (10) It follows from Eq.(9) that J.tk and Vk are tied through (l+'T}k)vk = l+'T}kJ.tk If we fix J.tk (and thus Vk) instead of treating it as an optimization variable, it would amount to a standard SVM regression problem with {-I, + I} outputs, the width of the asymmetric f-insensitive tube being J.tk-Vk = (~~~;)' This remark makes it possible for the reader to verify the results we reported on MNIST. Vsing the publicly available SVM software SVMtorch [1] with C = 10 and f = 0.1 as the width of the f-tube yields a 10-class error rate of 1.15% while the best performance using SVMtorch in classification mode is 1.3% (in both cases, we use Gaussian kernels with parameter (J = 1650). An explicit minimization on J.tk requires to add to the standard SVM regression problem the following constraint over the Lagrange multipliers (we use the same notation as in [9]) : Yi= l Yi=- l Yi= l Yi=- l Note that we still have the standard regression constraint I: ai = I: ai This has not been implemented yet , as we question the pertinence of the ~; slack variables for XVMs. Experiments with SVMtorch on a variety of tasks where non-zero slacks are required to achieve optimal performance (Reuters, VCI/Forest, VCI/Breast cancer) have not shown significant improvement using the regression mode while we vary the width of the f-tube. Many experiments on SVMs have reported that removing the outliers often gives efficient and sparse solutions. The early stopping heuristics that we have presented for XVMs suggest strategies to avoid learning (or to unlearn) the outliers, and this is the approach we are currently exploring. 7 Concluding Remarks This paper shows that large margin classification on extrapolated data is equivalent to the addition of the minimization of a second external margin to the standard SVM approach. The associated optimization problem is solved efficiently with convex hull distance minimization algorithms. A 1 % error rate is obtained on the MNIST dataset: it is the lowest ever obtained without a-priori knowledge about the data. We are currently trying to identify what other types of dataset show similar gains over SVMs, to determine how dependent XVM performance is on the facts that the data is separable or has invariance properties. We have only explored a few among the many variations the XVM models and algorithms allow , and a justification of why and when they generalize would help model selection. Geometry-based algorithms that handle potential outliers are also under investigation. Learning Theory bounds that would be a function of both the margin and some form of variance of the data would be necessary to predict XVM generalization and allow us to also consider the extrapolation factor 'TJ as an optimization variable. References [1] R. Collobert and S. Bengio. Support vector machines for large-scale regression problems. Technical Report IDIAP-RR-00-17, IDIAP, 2000. [2] C. Cortes and V. Vapnik. Support vector networks. Machine Learning, 20:1- 25 , 1995. [3] D. Crisp and C.J.C. Burges. A geometric interpretation of v-SVM classifiers. In Advances in Neural Information Processing Systems 12, S. A. Solla, T. K. Leen, K.-R. Mller, eds, Cambridge, MA, 2000. MIT Press. [4] D. DeCoste and B. Schoelkopf. Training invariant support vector machines. Machine Learning, special issue on Support Vector Machines and Methods, 200l. [5] S.S. Keerthi, S.K. Shevade, C. Bhattacharyya, and K.R.K. Murthy. A fast iterative nearest point algorithm for support vector machine classifier design. IEEE transactions on neural networks, 11(1):124 - 136, jan 2000. [6] A. Kowalczyk. Maximal margin perceptron. In Advances in Large Margin Classifiers, Smola, Bartlett, Schlkopf, and Schuurmans, editors, Cambridge, MA, 2000. MIT Press. [7] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition. proceedings of the IEEE, 86(11), 1998. [8] J. Platt, N. Christianini, and J. Shawe-Taylor. Large margin dags for multiclass classification. In Advances in Neural Information Processing Systems 12, S. A. Solla, T. K. Leen, K.-R. Mller, eds, Cambridge, MA, 2000. MIT Press. [9] V. N. Vapnik. Statistical Learning Theory. John Wiley & Sons, New-York, 1998.
2037 |@word trial:1 polynomial:5 stronger:1 tried:1 decomposition:1 concise:1 harder:1 configuration:2 contains:4 att:1 disparity:1 document:1 bhattacharyya:1 err:10 com:1 nt:2 surprising:1 yet:1 must:1 john:1 numerical:1 happen:1 reproducible:1 extrapolating:1 treating:1 v:3 leaf:1 fewer:1 jkj:1 ron:1 location:1 hyperplanes:1 simpler:2 along:3 consists:1 combine:2 manner:1 inter:2 expected:1 behavior:1 eck:2 decoste:1 becomes:1 moreover:1 notation:2 maximizes:2 lowest:3 what:2 kg:2 ehk:1 developed:1 nj:1 every:2 act:1 voting:1 tackle:1 classifier:19 k2:1 platt:1 ly:1 before:3 positive:1 treat:1 despite:1 interpolation:2 chose:1 studied:1 suggests:1 ease:1 range:2 statistically:1 lecun:1 testing:3 digit:2 procedure:1 jan:1 significantly:4 projection:1 word:1 suggest:3 get:1 onto:1 undesirable:2 selection:1 crisp:1 equivalent:3 map:2 lagrangian:1 center:2 maximizing:1 convex:14 wit:1 his:2 handle:2 traditionally:2 variation:1 justification:1 target:1 suppose:1 play:1 us:1 recognition:1 jk:5 asymmetric:1 predicts:2 database:2 observed:1 lix:1 role:1 solved:2 schoelkopf:1 solla:2 yk:10 intuition:1 discriminates:1 mu:1 ideally:1 trained:2 segment:1 crit:1 upon:1 kxj:2 differently:1 train:1 separated:1 fast:1 kp:6 outside:2 choosing:1 heuristic:4 larger:1 solve:1 say:1 relax:1 unseen:1 transductive:1 final:1 rr:1 propose:1 maximal:3 achieve:1 description:1 vsing:1 convergence:5 double:4 incremental:1 converges:1 tk:7 spent:1 help:1 measured:1 ij:2 nearest:6 eq:8 implemented:2 idiap:2 direction:4 correct:1 hull:14 require:7 fix:1 generalization:4 preliminary:1 investigation:2 elementary:1 extension:2 exploring:1 c_:1 around:1 considered:2 exp:3 k3:3 scope:2 predict:1 mller:2 vary:1 early:1 smallest:1 estimation:1 label:3 currently:2 largest:1 minimization:5 mit:3 always:1 gaussian:3 ck:6 rather:1 avoid:1 vk:19 improvement:1 likelihood:1 laurel:1 ave:1 spared:1 baseline:1 sense:1 inference:1 dependent:1 stopping:6 typically:1 hovers:1 quasi:1 pixel:3 issue:2 dual:6 aforementioned:1 among:4 classification:4 priori:2 constrained:1 special:1 once:1 look:3 minimized:1 others:2 report:1 few:3 geometry:1 keerthi:1 vsr:7 intra:2 light:1 primal:4 pc:2 tj:1 necessary:2 respective:1 f3ixi:1 tree:1 taylor:1 stopped:1 instance:2 soft:1 downside:1 maximization:1 introducing:1 reported:6 sv:10 again:2 tube:3 containing:1 choose:1 worse:1 external:4 account:3 potential:1 b2:2 caused:1 depends:1 vi:5 collobert:1 later:1 extrapolation:10 lab:1 observer:1 analyze:1 reached:1 contribution:1 publicly:1 variance:5 percept:1 maximized:2 correspond:1 yield:3 efficiently:1 identify:1 generalize:2 handwritten:1 critically:1 schlkopf:1 researcher:1 published:1 murthy:1 explain:1 reach:1 ed:2 associated:1 stop:8 gain:3 dataset:2 knowledge:3 higher:1 follow:1 improved:2 formulation:1 leen:2 just:1 smola:1 shevade:1 working:1 fallon:1 vci:2 mode:2 normalized:5 true:3 multiplier:6 verify:2 contain:2 read:1 width:5 unnormalized:2 criterion:4 trying:1 ranging:1 common:1 unlearn:1 jp:1 insensitive:1 belong:2 interpretation:4 he:5 extend:1 interpret:1 significant:2 cambridge:3 imposing:1 ai:2 dag:1 fk:4 similarly:2 shawe:1 dj:1 had:1 similarity:1 longer:1 surface:1 patrick:1 add:2 optimizing:1 binary:1 success:1 yi:6 minimum:4 additional:1 recognized:1 determine:2 maximize:3 converge:1 ii:4 full:1 desirable:1 ykw:2 rj:2 technical:1 characterized:1 x28:1 offer:2 long:1 y_:1 lin:1 mle:2 equally:1 regression:5 breast:1 iteration:1 kernel:13 represent:1 normalization:1 justified:1 whereas:1 addition:1 rest:1 seem:1 call:1 unitary:2 bengio:2 easy:2 variety:2 xj:4 gave:1 escaping:1 f3i:1 reduce:1 idea:1 haffner:3 y112:1 multiclass:3 bartlett:1 tain:1 effort:1 wo:1 york:1 cause:2 svs:1 remark:2 abies:1 amount:1 extensively:1 svms:19 reduced:1 outperform:1 dotted:1 sign:1 per:2 threshold:3 k4:6 christianini:1 sum:1 mystery:1 reader:1 decide:1 separation:4 decision:2 prefer:1 ob:1 bound:1 simplification:1 lec:1 lyk:1 constraint:19 software:1 interpolated:1 min:3 concluding:1 separable:4 o_:1 according:1 combination:2 poor:1 belonging:1 slightly:2 son:1 sam:1 dagsvm:1 wi:1 happens:1 outlier:4 invariant:1 taken:1 equation:1 previously:1 vs1:3 slack:5 count:1 end:1 available:1 k5:4 observe:1 ocr:2 enforce:1 kowalczyk:1 alternative:1 hat:1 iix:2 recombination:1 bl:1 added:1 question:1 strategy:1 usual:1 traditional:3 ow:1 win:1 gradient:1 distance:4 pertinence:1 separating:1 discriminant:3 trivial:1 toward:1 furthest:2 rjk:4 index:2 ratio:7 minimizing:2 pie:1 relate:1 negative:2 kfk:1 design:1 perform:3 extended:1 ever:2 pair:7 required:1 beyond:3 impair:1 below:1 max:3 explanation:1 rely:1 hybrid:7 improve:1 lk:7 kj:2 geometric:4 asymptotic:1 fully:1 bear:1 interesting:3 proportional:1 squeezing:1 validation:2 degree:2 offered:1 sufficient:1 consistent:1 editor:1 story:1 share:1 cancer:1 extrapolated:16 keeping:1 side:1 allow:5 burges:1 perceptron:1 sparse:1 plain:1 computes:1 commonly:1 far:1 transaction:1 implicitly:1 keep:1 kkt:1 b1:2 xi:14 search:2 iterative:3 why:2 table:11 ku:1 forest:1 schuurmans:1 bottou:1 did:3 reuters:1 fig:3 retrain:1 wiley:1 explicit:1 tied:1 down:1 theorem:1 removing:1 list:1 explored:1 svm:11 cortes:2 exists:1 ilwll:1 mnist:13 consist:1 adding:3 vapnik:3 margin:39 nk:6 gap:1 easier:1 subtract:1 ilx:1 lagrange:5 expressed:1 corresponds:3 chance:1 ma:3 replace:1 fisher:2 man:1 shared:1 absence:1 reducing:1 hyperplane:2 called:2 total:3 duality:7 invariance:3 partly:1 internal:3 support:20 violated:1 svmtorch:3 phenomenon:1
1,138
2,038
Spike timing and the coding of naturalistic sounds in a central auditory area of songbirds      Brian D. Wright, Kamal Sen, William Bialek and Allison J. Doupe Sloan?Swartz Center for Theoretical Neurobiology  Departments of Physiology and Psychiatry University of California at San Francisco, San Francisco, California 94143?0444  NEC Research Institute, 4 Independence Way, Princeton, New Jersey 08540 Department of Physics, Princeton University, Princeton, New Jersey 08544 bdwright/kamal/ajd @phy.ucsf.edu, wbialek@princeton.edu Abstract In nature, animals encounter high dimensional sensory stimuli that have complex statistical and dynamical structure. Attempts to study the neural coding of these natural signals face challenges both in the selection of the signal ensemble and in the analysis of the resulting neural responses. For zebra finches, naturalistic stimuli can be defined as sounds that they encounter in a colony of conspecific birds. We assembled an ensemble of these sounds by recording groups of 10-40 zebra finches, and then analyzed the response of single neurons in the songbird central auditory area (field L) to continuous playback of long segments from this ensemble. Following methods developed in the fly visual system, we measured the information that spike trains provide about the acoustic stimulus without any assumptions about which features of the stimulus are relevant. Preliminary results indicate that large amounts of information are carried by spike timing, with roughly half of the information accessible only at time resolutions better than 10 ms; additional information is still being revealed as time resolution is improved to 2 ms. Information can be decomposed into that carried by the locking of individual spikes to the stimulus (or modulations of spike rate) vs. that carried by timing in spike patterns. Initial results show that in field L, temporal patterns give at least  % extra information. Thus, single central auditory neurons can provide an informative representation of naturalistic sounds, in which spike timing may play a significant role. 1 Introduction Nearly fifty years ago, Barlow [1] and Attneave [2] suggested that the brain may construct a neural code that provides an efficient representation for the sensory stimuli that occur in the natural world. Slightly earlier, MacKay and McCulloch [3] emphasized that neurons that could make use of spike timing?rather than a coarser ?rate code??would have available a vastly larger capacity to convey information, although they left open the question of whether this capacity is used efficiently. Theories for timing codes and efficient representation have been discussed extensively, but the evidence for these attractive ideas remains tenuous. A real attack on these issues requires (at least) that we actually measure the information content and efficiency of the neural code under stimulus conditions that approximate the natural ones. In practice, constructing an ensemble of ?natural? stimuli inevitably involves compromises, and the responses to such complex dynamic signals can be very difficult to analyze. At present the clearest evidence on efficiency and timing in the coding of naturalistic stimuli comes from central invertebrate neurons [4, 5] and from the sensory periphery [6, 7] and thalamus [8, 9] of vertebrates. The situation for central vertebrate brain areas is much less clear. Here we use the songbird auditory system as an accessible test case for these ideas. The set of songbird telencephalic auditory areas known as the field L complex is analogous to mammalian auditory cortex and contains neurons that are strongly driven by natural sounds, including the songs of birds of the same species (conspecifics) [10, 11, 12, 13]. We record from the zebra finch field L, using naturalistic stimuli that consist of recordings from groups of 10-40 conspecific birds. We find that single neurons in field L show robust and well modulated responses to playback of long segments from this song ensemble, and that we are able to maintain recordings of sufficient stability to collect the large data sets that are required for a model independent information theoretic analysis. Here we give a preliminary account of our experiments. 2 A naturalistic ensemble Auditory processing of complex sounds is critical for perception and communication in many species, including humans, but surprisingly little is known about how high level brain areas accomplish this task. Songbirds provide a useful model for tackling this issue, because each bird within a species produces a complex individualized acoustic signal known as a song, which reflects some innate information about the species? song as well as information learned from a ?tutor? in early life. In addition to learning their own song, birds use the acoustic information in songs of others to identify mates and group members, to discriminate neighbors from intruders, and to control their living space [14]. Consistent with how ethologically critical these functions are, songbirds have a large number of forebrain auditory areas with strong and increasingly specialized responses to songs [11, 15, 16]. The combination of a rich set of behaviorally relevant stimuli and a series of high-level auditory areas responsive to those sounds provides an opportunity to reveal general principles of central neural encoding of complex sensory stimuli. Many prior studies have chosen to study neural responses to individual songs or altered versions thereof. In order to make the sounds studied increasingly complex and natural, we have made recordings of the sounds encountered by birds in our colony of zebra finches. To generate the sound ensemble that was used in this study we first created long records of the vocalizations of groups of 10-40 zebra finches in a soundproof acoustic chamber with a directional microphone above the bird cages. The group of birds generated a wide variety of vocalizations including songs and a variety of different types of calls. Segments of these sounds were then joined to cre ate the sounds presented in the experiment. One of the segments that was presented ( sec) was repeated in alternation with different segments.  We recorded the neural responses in field L of one of the birds from the group to the ensemble of natural sounds played back through a speaker, at an intensity approximately equal to that in the colony recording. This bird was lightly anesthetized with urethane. We used a single electrode to record the neural response waveforms and sorted single units offline. Further details concerning experimental techniques can be found in Ref. [13]. A D 500 ms B 50 Hz C Figure 1: A. Spike raster of 4 seconds of the responses of a single neuron in field L to a 30 second segment of a natural sound ensemble of zebra finch sounds. The stimulus was repeated 80 times. B. Peri-stimulus time histogram (PSTH) with 1 ms bins. C. Sound pressure waveform for the natural sound ensemble. D. Blowup of segment shown in the box in A. The scale bar is 50 ms. 3 Information in spike sequences The auditory telencephalon of birds consists of a set of areas known as the field L complex, which receive input from the auditory thalamus and project to increasingly selective auditory areas such as NCM, cHV and NIf [12, 17] and ultimately to the brain areas specialized for the bird?s own song. Field L neurons respond to simple stimuli such as tone bursts, and are organized in a roughly tonotopic fashion [18], but also respond robustly to many complex sounds, including songs. Figure 1 shows 4 seconds of the responses of a cell in field L to repeated presentations of a 30 sec segment from the natural ensemble described above. Averaging over presentations, we see that spike rates are well modulated. Looking at the responses on a finer time resolution we see that aspects of the spike train are reproducible  ms time scale. This encourages us to measure the information content of on at least a these responses over a range of time scales, down to millisecond resolution.    Our approach to estimating the information content of spike trains follows Ref. [4]. At some time (defined relative to the repeating stimulus) we open a window of size to look at the response. Within this window we discretize the spike arrival times with resolution so that the response becomes a ?word? with letters. If the time resolution is very small, the allowed letters are only 1 and 0, but as becomes larger one must keep track of multiple spikes within each bin. Examining the whole experiment, we sample       40 Total Entropy Noise Entropy Mutual Info 35 Information Rate (bits/sec) 30 25 20 15 10 5 0 0 0.01 0.02 0.03 0.04 0.05 0.06 1/Nrepeats   Figure 2: Mutual information rate for the spike train is shown as a function of data size for  ms.  ms and  the probability distribution of words, , and the entropy of this distribution sets the capacity of the code to convey information about the stimulus:          !" #%$'&)(+* (1) where the notation reminds us that the entropy depends both on the size of the words that we consider and on the time resolution with which we classify the responses. We can think of this entropy as measuring the size of the neuron?s vocabulary. , Because the whole experiment contributes to defining the vocabulary size, estimating the distribution and hence the total entropy is not significantly limited by the problems of finite sample size. This can be seen in Fig. 2 in the stability of the total entropy with changing the number of repeats used in the analysis. Here we show the total entropy as a rate in bits per second by dividing the entropy by the time window .  While the capacity of the code is limited by the total entropy, to convey information particular words in the vocabulary must be associated, more or less reliably, with particular stimulus features. If we look at one time relative to the (long) stimulus, and examine the words generated on repeated presentations, we sample the conditional distribution . This distribution has an entropy that quantifies the noise in the response at time , and averaging over all times we obtain the average noise entropy,  10 )2 34  5  7689   "/.  :';   "/.  =<,>?#%$'&)(+*  -/.  (2) > where  indicates a time average (in general,   denotes an average over the variable ). Technically, the above average should be an average over stimuli , however, for a sufficiently long and rich stimulus, the ensemble average over can be replaced by a time average. For the noise entropy, the problem of sampling is much more severe, since each distribution is estimated from a number of examples given by the number of repeats. Still, as shown in Fig. 2, we find that the dependence of our estimate on sample size is simple and regular; specifically, we find   /.   5    4 4  3  5      4  4  3   (3) This is what we expect for any entropy estimate if the distribution is well sampled, and if we make stronger assumptions about the sampling process (independence of trials etc.) we can even estimate the correction coefficient  [19]. In systems where much larger data sets are available this extrapolation procedure has been checked, and the observation of a good fit to Eq. (3) is a strong indication that larger sample sizes will be consistent  with ; further, this extrapolation can be tested against bounds on the entropy that are derived from more robust quantities [4]. Most importantly, failure to observe Eq. (3) means that we are in a regime where sampling is not sufficient to draw reliable conclusions without more sophisticated arguments, and we exclude these regions of and from our discussion.      5    Ideally, to measure the spike train total and noise entropy rates, we want to go to the limit of infinite word duration. A true entropy is extensive, which here means that it grows linearly with spike train word duration , so that the entropy rate  is constant. For finite word duration however, words sampled at neighboring times will have correlations between them due, in part, to correlations in the stimulus (for birdsong these stimulus autocorrelation time scales can extend up to   ms). Since the word samples are not completely independent, the raw entropy rate is an overestimate of the true entropy rate. The effect is larger for smaller word duration and the leading dependence of the raw estimate is    5              * (4) where   and we have already taken the infinite data size limit. We cannot directly take the large limit, since for large word lengths we eventually reach a data sampling limit beyond which we are unable to reliably compute the word distributions. On the other hand, if there is a range of for which the distributions are sufficiently well sampled, the behavior in Eq. (4) should be observed and can be used to extrapolate to infinite word size [4]. We have checked that our data shows this behavior and that it sets in for word sizes below the limit where the data sampling problem occurs. For example, in the case of  ms, it applies for below the limit of   ms (above this we the noise entropy, for run into sampling problems). The total entropy estimate is nearly perfectly extensive.     Finally, we combine estimates of total and noise entropies to obtain the information that words carry about the sensory stimulus,        5   0 2 3 4  5  #%$'&)(   (5) Figure 2 shows the total and noise entropy rates as well as the mutual information rate for  ms and time resolution  ms. The error bars on the raw a time window entropy and information rates were estimated to be approximately !    bits/sec using a simple bootstrap procedure over the repeated trials. The extrapolation to infinite data size is shown for the mutual information rate estimate (error bars in the extrapolated values will be "#!    bits/sec) and is consistent with the prediction of Eq. (3). Since the total entropy is nearly extensive and the noise entropy rate decreases with word duration due to subextensive corrections as described above, the mutual information rate shown in Fig. 2 grows with word duration. We find that there is an upward change in the mutual information 5 Information Rate (bits/sec) 4.5 Spike Train 4 Independent Events 3.5 3 2.5 2 1.5 0 5 10 15 ?? (ms) 20 25 30 35     ms) and single spike events as a Figure 3: Information rates for the spike train ( function of time resolution of the spike rasters, corrected for finite data size effects.       ms) of  ms and rate (computed at %, in the large limit. For  ms that is in the simplicity in the following, we shall look at a fixed word duration well-sampled region for all time resolutions considered.   The mutual information rate measures the rate at which the spike train removes uncertainty about the stimulus. However, the mutual information estimate does not depend on identifying either the relevant features of the stimulus or the relevant features of the response, which is crucial in analyzing the response to such complex stimuli. In this sense, our estimates of information transmission and efficiency are independent of any model for the code, and provide a benchmark against which such models could be tested.     One way to look at the information results is to fix our time window and ask what happens as we change our time resolution . When , the ?word? describing the response is nothing but the number of spikes in the window, so we have a rate or counting code. As we decrease , we gradually distinguish more and more detail in the   ms arrangement of spikes in the window. We chose a range of values from  in our analyses to cover previously observed response windows for field L neurons and to  ms) of individual song syllables or notes. probe the behaviorally relevant time scale (  For ms, we show the results (extrapolated to infinite data size) in the upper curve of Fig. 3. The spike train mutual information shows a clear increase as the timing resolution is improved. In addition, Fig. 3 shows that roughly half of the information is accessible at time resolutions better than  ms and additional information is still being revealed as time resolution is improved to 2 ms.     4 Information in rate modulation  Knowing the mutual information between the stimulus and the spike train (defined in the window ), we would like to ask whether this can be accounted for by the information in single spike events or whether there is some additional information conveyed by the patterns of spikes. In the latter case, we have precisely what we mean by a temporal or timing code: there is information beyond that attributable to the probability of single spike events occurring at time relative to the onset of the stimulus. By event at time , we mean that the event occurs between time and time , where is the resolution at which we are looking at the spike train. This probability is simply proportional to the firing rate (or peri-stimulus time histogram (PSTH)) at time normalized by the mean firing rate   . Specifically if the duration of each repeated trial is  we have         .   spk @              4 4   4  4  * (6)      4  spk @  . 4  where denotes the stimulus history ( " ). The probability of a spike event at ,  a priori of knowing the stimulus history, is flat: spk @ . Thus, the  mutual information between the stimulus and the single spike events is [20]:  (%$      spk @    6           -< > #%$'&)(+*    (7) where is the PSTH binned to resolution and the stimulus average in the first expression is replaced by a time average in the second (as discussed in the calculation of the noise entropy in spike train words in the previous section). We find that this information  ms. Supposing that the individual spike events are inis approximately bit for dependent (i.e. no intrinsic spike train correlations), the information rate in single spike events is obtained by multiplying the mutual information per spike (Eq. 7) by the mean   Hz). This gives an upper bound to the single spike event firing rate of the neuron ( contribution to the information rate and is shown in the lower curve of Fig. 3 (error bars are again " !    bits/sec). Comparing with the spike train information (upper curve), we  ms, there is at least  % of the total information in see that at a resolution of the spike train that cannot be attributable to single spike events. Thus there is some pattern of spikes that is contributing synergistically to the mutual information. The fact discussed, in the previous section, that the spike train information rate grows subextensively with the the word duration out to the point where data sampling becomes problematic is further confirmation of the synergy from spike patterns. Thus we have shown model-independent evidence for a temporal code in the neural responses.   8 5 Conclusion Until now, few experiments on neural responses in high level, central vertebrate brain areas have measured the information that these responses provide about dynamic, naturalistic sensory signals. As emphasized in earlier work on invertebrate systems, information theoretic approaches have the advantage that they require no assumptions about the features of the stimulus to which neurons respond. Using this method in the songbird auditory forebrain, we found that patterns of spikes seem to be special events in the neural code of these neurons, since they carry more information than expected by adding up the contributions of individual spikes. It remains to be determined what these spike patterns are, what stimulus features they may encode, and what mechanisms may be responsible for reading such codes at even higher levels of processing. Acknowledgments Work at UCSF was supported by grants from the NIH (NS34835) and the Sloan-Swartz Center for Theoretical Neurobiology. BDW and KS supported by NRSA grants from the NIDCD. We thank Katrin Schenk and Robert Liu for useful discussions. References 1. Barlow, H.B. (1961). Possible principles underlying the transformation of sensory messages. In Sensory Communication, W.A. Rosenblith, ed., pp. 217?234 (MIT Press, Cambridge, MA). 2. Attneave, F. (1954). Some informational aspects of visual perception. Psychol. Rev. 61, 183?193. 3. MacKay, D. and McCulloch, W.S. (1952). The limiting information capacity of a neuronal link. Bull. Math. Biophys. 14, 127?135. 4. Strong, S.P., Koberle, R., de Ruyter van Steveninck, R. and Bialek, W. (1998). Entropy and information in neural spike trains, Phys. Rev. Lett. 80, 197?200. 5. Lewen, G.D., Bialek, W. and de Ruyter van Steveninck, R.R. (2001). Neural coding of naturalistic motion stimuli. Network 12, 317?329. 6. Rieke, F., Bodnar, D.A. and Bialek, W. (1995). Naturalistic stimuli increase the rate and efficiency of information transmission by primary auditory afferents. Proc. R. Soc. Lond. B 262, 259?265. 7. Berry II, M.J., Warland, D.K. and Meister, M. (1997). The structure and precision of retinal spike trains. Proc. Nat. Acad. Sci. (USA) 94, 5411?5416. 8. Reinagel, P. and Reid, R.C. (2000). Temporal coding of visual information in the thalamus. J. Neurosci. 20, 5392?5400. 9. Liu, R.C., Tzonev, S., Rebrik, S. and Miller, K.D. (2001). Variability and information in a neural code of the cat lateral geniculate nucleus. J. Neurophysiol. 86, 2789?2806. 10. Scheich, H., Langner, G. and Bonke, D. (1979). Responsiveness of units in the auditory neostriatum of the guinea fowl (Numida meleagris) to species-specific calls and synthetic stimuli II. Discrimination of Iambus-Like Calls. J. Comp. Physiol. A 132, 257?276. 11. Lewicki, M.S. and Arthur, B.J. (1996). Hierarchical organization of auditory temporal context sensitivity. J. Neurosci. 16(21), 6987?6998. 12. Janata, P. and Margoliash, D. (1999). Gradual emergence of song selectivity in sensorimotor structures of the male zebra finch song system. J. Neurosci. 19(12), 5108?5118. 13. Theunissen, F.E., Sen, K. and Doupe, A.J. (2000). Spectral temporal receptive fields of nonlinear auditory neurons obtained using natural sounds. J. Neurosci. 20(6), 2315?2331. 14. Searcy, W.A. and Nowicki, S. (1999). In The Design of Animal Communication, M.D. Hauser and M. Konishi, eds., pp. 577?595 (MIT Press, Cambridge, MA). 15. Margoliash, D. (1983). Acoustic parameters underlying the responses of song-specific neurons in the white-crowned sparrow. J. Neurosci. 3(5), 1039?1057. 16. Sen, K., Theunissen, F.E. and Doupe, A.J. (2001). Feature analysis of natural sounds in the songbird auditory forebrain. J. Neurophysiol. 86, 1445?1458. 17. Stripling, R., Kruse, A.A. and Clayton, D.F. (2001). Development of song responses in the zebra finch caudomedial neostriatum: role of genomic and electrophysiological activities. J. Neurobiol. 48, 163?180. 18. Zaretsky, M.D. and Konishi, M. (1976). Tonotopic organization in the avian telencephalon. Brain Res. 111, 167?171. 19. Treves, A. and Panzeri, S. (1995). The upward bias in measures of information derived from limited data samples. Neural Comput., 7, 399?407. 20. Brenner, N., Strong, S., Koberle, R. and Bialek, W. (2000). Synergy in a neural code, Neural Comput. 12, 1531?1552.
2038 |@word trial:3 version:1 stronger:1 open:2 gradual:1 pressure:1 carry:2 initial:1 synergistically:1 contains:1 series:1 liu:2 phy:1 comparing:1 tackling:1 must:2 physiol:1 informative:1 remove:1 reproducible:1 v:1 discrimination:1 half:2 ajd:1 tone:1 rebrik:1 record:3 provides:2 math:1 psth:3 attack:1 burst:1 consists:1 combine:1 autocorrelation:1 expected:1 blowup:1 behavior:2 roughly:3 examine:1 brain:6 informational:1 decomposed:1 little:1 window:9 vertebrate:3 chv:1 project:1 estimating:2 becomes:3 notation:1 underlying:2 mcculloch:2 what:6 neurobiol:1 developed:1 transformation:1 temporal:6 control:1 unit:2 grant:2 reid:1 overestimate:1 timing:9 limit:7 acad:1 encoding:1 analyzing:1 firing:3 modulation:2 approximately:3 chose:1 bird:12 studied:1 k:1 collect:1 limited:3 range:3 steveninck:2 acknowledgment:1 responsible:1 practice:1 bootstrap:1 procedure:2 area:11 physiology:1 significantly:1 word:22 regular:1 naturalistic:9 cannot:2 selection:1 context:1 center:2 go:1 duration:9 resolution:17 simplicity:1 identifying:1 reinagel:1 importantly:1 stability:2 rieke:1 konishi:2 analogous:1 limiting:1 margoliash:2 play:1 mammalian:1 coarser:1 theunissen:2 observed:2 role:2 fly:1 region:2 decrease:2 locking:1 ideally:1 dynamic:2 ultimately:1 depend:1 segment:8 compromise:1 technically:1 efficiency:4 completely:1 neurophysiol:2 spk:4 langner:1 jersey:2 cat:1 train:19 wbialek:1 larger:5 tested:2 think:1 emergence:1 vocalization:2 sequence:1 indication:1 advantage:1 sen:3 neighboring:1 relevant:5 electrode:1 transmission:2 produce:1 avian:1 colony:3 measured:2 eq:5 strong:4 dividing:1 soc:1 involves:1 indicate:1 come:1 waveform:2 human:1 bin:2 require:1 fix:1 preliminary:2 brian:1 correction:2 sufficiently:2 considered:1 wright:1 panzeri:1 early:1 proc:2 geniculate:1 reflects:1 mit:2 behaviorally:2 genomic:1 playback:2 rather:1 encode:1 derived:2 indicates:1 psychiatry:1 sense:1 dependent:1 selective:1 upward:2 issue:2 priori:1 development:1 animal:2 special:1 mackay:2 mutual:13 field:12 construct:1 equal:1 sampling:7 look:4 nearly:3 kamal:2 fowl:1 others:1 stimulus:39 few:1 individual:5 replaced:2 conspecific:3 william:1 maintain:1 attempt:1 organization:2 message:1 severe:1 male:1 allison:1 analyzed:1 arthur:1 re:1 theoretical:2 classify:1 earlier:2 cover:1 measuring:1 bull:1 examining:1 hauser:1 accomplish:1 finch:8 synthetic:1 cre:1 peri:2 sensitivity:1 accessible:3 physic:1 vastly:1 central:7 recorded:1 again:1 leading:1 account:1 exclude:1 de:2 retinal:1 coding:5 sec:7 coefficient:1 sloan:2 afferent:1 depends:1 onset:1 extrapolation:3 analyze:1 contribution:2 nidcd:1 efficiently:1 ensemble:12 miller:1 identify:1 directional:1 raw:3 multiplying:1 comp:1 finer:1 ago:1 history:2 reach:1 phys:1 rosenblith:1 checked:2 ed:2 against:2 raster:2 failure:1 sensorimotor:1 pp:2 clearest:1 thereof:1 attneave:2 associated:1 sampled:4 auditory:18 ask:2 electrophysiological:1 organized:1 sophisticated:1 actually:1 back:1 higher:1 response:25 improved:3 box:1 strongly:1 nif:1 correlation:3 hand:1 until:1 nonlinear:1 cage:1 reveal:1 grows:3 innate:1 usa:1 effect:2 normalized:1 true:2 barlow:2 hence:1 nowicki:1 reminds:1 attractive:1 white:1 encourages:1 songbird:8 speaker:1 m:25 theoretic:2 motion:1 nih:1 specialized:2 discussed:3 extend:1 significant:1 cambridge:2 zebra:8 crowned:1 cortex:1 etc:1 own:2 driven:1 periphery:1 selectivity:1 life:1 alternation:1 seen:1 responsiveness:1 additional:3 tzonev:1 swartz:2 kruse:1 living:1 signal:5 ii:2 multiple:1 sound:20 thalamus:3 sparrow:1 calculation:1 long:5 concerning:1 prediction:1 histogram:2 cell:1 receive:1 addition:2 want:1 crucial:1 extra:1 fifty:1 recording:5 hz:2 supposing:1 member:1 seem:1 call:3 counting:1 revealed:2 variety:2 independence:2 fit:1 perfectly:1 idea:2 knowing:2 whether:3 expression:1 birdsong:1 song:16 useful:2 clear:2 amount:1 repeating:1 extensively:1 generate:1 problematic:1 millisecond:1 estimated:2 track:1 per:2 shall:1 group:6 changing:1 year:1 run:1 letter:2 uncertainty:1 respond:3 intruder:1 draw:1 bit:7 bound:2 played:1 distinguish:1 syllable:1 encountered:1 activity:1 binned:1 occur:1 precisely:1 flat:1 katrin:1 invertebrate:2 lightly:1 aspect:2 argument:1 lond:1 department:2 combination:1 tonotopic:2 ate:1 slightly:1 increasingly:3 smaller:1 rev:2 happens:1 gradually:1 taken:1 remains:2 forebrain:3 describing:1 eventually:1 previously:1 mechanism:1 meister:1 available:2 probe:1 observe:1 hierarchical:1 spectral:1 chamber:1 responsive:1 robustly:1 encounter:2 denotes:2 opportunity:1 warland:1 tutor:1 question:1 quantity:1 spike:50 already:1 occurs:2 arrangement:1 dependence:2 primary:1 receptive:1 bialek:5 individualized:1 unable:1 thank:1 capacity:5 link:1 sci:1 lateral:1 code:14 length:1 difficult:1 robert:1 info:1 design:1 reliably:2 discretize:1 ethologically:1 neuron:15 observation:1 upper:3 benchmark:1 mate:1 finite:3 neostriatum:2 inevitably:1 nrsa:1 situation:1 neurobiology:2 communication:3 looking:2 defining:1 variability:1 intensity:1 treves:1 clayton:1 required:1 extensive:3 california:2 acoustic:5 learned:1 assembled:1 able:1 suggested:1 bar:4 dynamical:1 pattern:7 perception:2 beyond:2 below:2 regime:1 reading:1 challenge:1 including:4 reliable:1 critical:2 event:13 natural:12 altered:1 created:1 carried:3 psychol:1 koberle:2 prior:1 lewen:1 berry:1 contributing:1 relative:3 expect:1 proportional:1 nucleus:1 conveyed:1 sufficient:2 consistent:3 principle:2 extrapolated:2 surprisingly:1 repeat:2 accounted:1 supported:2 guinea:1 offline:1 bias:1 institute:1 neighbor:1 wide:1 face:1 anesthetized:1 van:2 curve:3 lett:1 vocabulary:3 world:1 rich:2 sensory:8 made:1 san:2 approximate:1 tenuous:1 keep:1 synergy:2 francisco:2 continuous:1 quantifies:1 nature:1 robust:2 confirmation:1 ruyter:2 contributes:1 complex:10 constructing:1 linearly:1 neurosci:5 whole:2 noise:10 arrival:1 nothing:1 repeated:6 ref:2 convey:3 allowed:1 neuronal:1 fig:6 fashion:1 attributable:2 precision:1 comput:2 down:1 specific:2 emphasized:2 evidence:3 consist:1 intrinsic:1 adding:1 nec:1 nat:1 occurring:1 biophys:1 entropy:29 simply:1 visual:3 ncm:1 joined:1 lewicki:1 applies:1 ma:2 conditional:1 sorted:1 presentation:3 brenner:1 content:3 change:2 specifically:2 infinite:5 corrected:1 determined:1 averaging:2 microphone:1 total:11 specie:5 discriminate:1 experimental:1 doupe:3 latter:1 modulated:2 ucsf:2 princeton:4 bodnar:1 extrapolate:1
1,139
2,039
ADynamic HMM for On-line Segmentation of Sequential Data Jens Kohlmorgen* Fraunhofer FIRST.IDA Kekulestr. 7 12489 Berlin, Germany Steven Lemm Fraunhofer FIRST.IDA Kekulestr. 7 12489 Berlin, Germany jek@first?fraunhofer.de lemm @first?fraunhofer.de Abstract We propose a novel method for the analysis of sequential data that exhibits an inherent mode switching. In particular, the data might be a non-stationary time series from a dynamical system that switches between multiple operating modes. Unlike other approaches, our method processes the data incrementally and without any training of internal parameters. We use an HMM with a dynamically changing number of states and an on-line variant of the Viterbi algorithm that performs an unsupervised segmentation and classification of the data on-the-fly, i.e. the method is able to process incoming data in real-time. The main idea of the approach is to track and segment changes of the probability density of the data in a sliding window on the incoming data stream. The usefulness of the algorithm is demonstrated by an application to a switching dynamical system. 1 Introduction Abrupt changes can occur in many different real-world systems like, for example, in speech, in climatological or industrial processes, in financial markets, and also in physiological signals (EEG/MEG). Methods for the analysis of time-varying dynamical systems are therefore an important issue in many application areas. In [12], we introduced the annealed competition of experts method for time series from nonlinear switching dynamics, related approaches were presented, e.g., in [2, 6, 9, 14]. For a brief review of some of these models see [5], a good introduction is given in [3]. We here present a different approach in two respects. First, the segmentation does not depend on the predictability of the system. Instead, we merely estimate the density distribution of the data and track its changes. This is particularly an improvement for systems where data is hard to predict, like, for example, EEG recordings [7] or financial data. Second, it is an on-line method. An incoming data stream is processed incrementally while keeping the computational effort limited by a fixed ? http://www.first.fraunhofer.de/..-.jek upper bound, i.e. the algorithm is able to perpetually segment and classify data streams with a fixed amount of memory and CPU resources. It is even possible to continuously monitor measured data in real-time, as long as the sampling rate is not too high.l The main reason for achieving a high on-line processing speed is the fact that the method, in contrast to the approaches above, does not involve any training, i.e. iterative adaptation of parameters. Instead, it optimizes the segmentation on-the-fly by means of dynamic programming [1], which thereby results in an automatic correction or fine-tuning of previously estimated segmentation bounds. 2 The segmentation algorithm We consider the problem of continuously segmenting a data stream on-line and simultaneously labeling the segments. The data stream is supposed to have a sequential or temporal structure as follows: it is supposed to consist of consecutive blocks of data in such a way that the data points in each block originate from the same underlying distribution. The segmentation task is to be performed in an unsupervised fashion, i.e. without any a-priori given labels or segmentation bounds. 2.1 Using pdfs as features for segmentation Consider Yl, Y2 , Y3, ... , with Yt E Rn, an incoming data stream to be analyzed. The sequence might have already passed a pre-processing step like filtering or subsampling, as long as this can be done on-the-fly in case of an on-line scenario. As a first step of further processing, it might then be useful to exploit an idea from dynamical systems theory and embed the data into a higher-dimensional space, which aims to reconstruct the state space of the underlying system, Xt = (Yt,Yt-n'" ,Yt -(m-l)r )' (1) The parameter m is called the embedding dimension and T is called the delay parameter of the embedding. The dimension of the vectors Xt thus is d = m n. The idea behind embedding is that the measured data might be a potentially non-linear projection of the systems state or phase space. In any case, an embedding in a higher-dimensional space might help to resolve structure in the data, a property which is exploited, e.g., in scatter plots. After the embedding step one might perform a sub-sampling of the embedded data in order to reduce the amount of data for real-time processing. 2 Next, we want to track the density distribution of the embedded data and therefore estimate the probability density function (pdf) in a sliding window of length W. We use a standard density estimator with multivariate Gaussian kernels [4] for this purpose, centered on the data points 3 in the window ~ }W -l { Xt-w w=o, () 1 ~l 1 Pt x = W ~ (27fa 2 )d/2 exp (x - Xt_w)2) ( - 2a 2 . (2) The kernel width a is a smoothing parameter and its value is important to obtain a good representation of the underlying distribution. We propose to choose a proportional to the mean distance of each Xt to its first d nearest neighbors, averaged over a sample set {xt}. 1 In our reported application we can process data at 1000 Hz (450 Hz including display) on a 1.33 GHz PC in MATLAB/C under Linux, which we expect is sufficient for a large number of applications. 2In that case, our further notation of time indices would refer to the subsampled data. 3We use if to denote a specific vector-valued point and x to denote a vector-valued variable. 2.2 Similarity of two pdfs Once we have sampled enough data points to compute the first pdf according to eq. (2), we can compute a new pdf with each new incoming data point. In order to quantify the difference between two such functions, f and g, we use the squared L2 -Norm, also called integrated squared error (ISE) , d(f, g) = J(f - g)2 dx , which can be calculated analytically if f and 9 are mixtures of Gaussians as in our case of pdfs estimated from data windows, (3) 2.3 The HMM in the off-line case Before we can discuss the on-line variant, it is necessary to introduce the HMM and the respective off-line algorithm first. For a given a data sequence, {X'dT=l' we can obtain the corresponding sequence of pdfs {Pt(X)}tES, S = {W, ... , T}, according to eq. (2). We now construct a hidden Markov model (HMM) where each of these pdfs is represented by a state s E S, with S being the set of states in the HMM. For each state s, we define a continuous observation probability distribution, - ( (X) I s-~ ) PPt 1 V 21f <; exp ( - d(Ps(X),Pt(x))) 22 <; ' (4) for observing a pdf Pt(x) in state s. Next, the initial state distribution {1f s LES of the HMM is given by the uniform distribution, 1fs = liN, with N = lSI being the number of states. Thus, each state is a-priori equally probable. The HMM transition matrix, A = (PijkjES, determines each probability to switch from a state Si to a state Sj. Our aim is to find a representation of the given sequence of pdfs in terms of a sequence of a small number of representative pdfs, that we call prototypes, which moreover exhibits only a small number of prototype changes. We therefore define A in such a way that transitions to the same state are k times more likely than transitions to any of the other states, _ { Pij - k+~-l 1 k+N - l ;ifi=J ;ifi-j.J (5) This completes the definition of our HMM. Note that this HMM has only two free parameters, k and <;. The well-known Viterbi algorithm [13] can now be applied to the above HMM in order to compute the optimal - i.e. the most likely - state sequence of prototype pdfs that might have generated the given sequence of pdfs. This state sequence represents the segmentation we are aiming at. We can compute the most likely state sequence more efficiently if we compute it in terms of costs, c = -log(p), instead of probabilities p, i.e. instead of computing the maximum of the likelihood function L , we compute the minimum of the cost function , -log(L), which yields the optimal state sequence as well. In this way we can replace products by sums and avoid numerical problems [13]. In addition to that, we can further simplify the computation for the special case of our particular HMM architecture, which finally results in the following algorithm: For each time step, t = w, ... ,T, we compute for all S E S the cost cs(t) of the optimal state sequence from W to t, subject to the constraint that it ends in state S at time t. We call these constrained optimal sequences c-paths and the unconstrained optimum 0* -path. The iteration can be formulated as follows, with ds,t being a short hand for d(ps(x)'pt(x)) and bs,s denoting the Kronecker delta function : Initialization, Vs E S: Cs(W) := (6) ds ,w, Induction, Vs E S: cs(t) := ds,t + min sES { cs(t - 1) + C (1- bs 's)}, for t = W + 1, ... , T, (7) Termination: 0* := (8) min { cs(T) } . sES The regularization constant C, which is given by C = 2C; 2 10g(k) and thus subsumes our two free HMM parameters, can be interpreted as transition cost for switching to a new state in the path. 4 The optimal prototype sequence with minimal costs 0* for the complete series of pdfs, which is determined in the last step, is obtained by logging and updating the c-paths for all states s during the iteration and finally choosing the one with minimal costs according to eq. (8). 2.4 The on-line algorithm In order to turn the above segmentation algorithm into an on-line algorithm, we must restrict the incremental update in eq. (7), such that it only uses pdfs (and therewith states) from past data points. We neglect at this stage that memory and CPU resources are limited. Suppose that we have already processed data up to T - 1. When a new data point YT arrives at time T, we can generate a new embedded vector XT (once we have sampled enough initial data points for the embedding), we have a new pdf pT(X) (once we have sampled enough embedded vectors Xt for the first pdf window), and thus we have given a new HMM state. We can also readily compute the distances between the new pdf and all the previous pdfs, dT,t, t < T, according to eq. (3). A similarly simple and straightforward update of the costs, the c-paths and the optimal state sequence is only possible, however, if we neglect to consider potential c-paths that would have contained the new pdf as a prototype in previous segments. In that case we can simply reuse the c-paths from T - 1. The on-line update at time T for these restricted paths, that we henceforth denote with a tilde, can be performed as follows: For T = W, we have cw(W) := o*(W) := dw,w = O. For T > W: 1. Compute the cost cT(T - 1) for the new state s For t = T - 1, compute w, ... , =T at time T - 1: 0 ift=W CT(t) :=dT,t+ { min{cT(t-1) ; o*(t-1)+C}: else (9) and update o*(t) := CT(t), if CT(t) < o*(t). (10) Here we use all previous optimal segmentations o*(t), so we don't need to keep the complete matrix (cs(t))S,tES and repeatedly compute the minimum 4We developed an algorithm that computes an appropriate value for the hyperparameter C from a sample set {it}. Due to the limited space we will present that algorithm in a forthcoming publication [8]. over all states. However, we must store and update the history of optimal segmentations 8* (t). 2. Update from T - 1 to T and compute cs(T) for all states s E S obtained so far, and also get 8*(T): For s = W, ... , T , compute cs(T) := ds,T + min {cs(T - 1); 8*(T - 1) + C} (11) and finally get the cost of the optimal path 8* (T) := min {cs(T)} . sES (12) As for the off-line case, the above algorithm only shows the update equations for the costs of the C- and 8* -paths. The associated state sequences must be logged simultaneously during the computation. Note that this can be done by just storing the sequence of switching points for each path. Moreover, we do not need to keep the full matrix (cs(t))s ,tES for the update, the most recent column is sufficient. So far we have presented the incremental version of the segmentation algorithm. This algorithm still needs an amount of memory and CPU time that is increasing with each new data point. In order to limit both resources to a fixed amount, we must remove old pdfs, i.e. old HMM states, at some point. We propose to do this by discarding all states with time indices smaller or equal to s each time the path associated with cs(T) in eq. (11) exhibits a switch back from a more recent state/pdf to the currently considered state s as a result of the min-operation in eq. (11). In the above algorithm this can simply be done by setting W := s + 1 in that case, which also allows us to discard the corresponding old cs(T)- and 8* (t)-paths, for all s::::: sand t < s. In addition, the "if t = W" initialization clause in eq. (9) must be ignored after the first such cut and the 8* (W - I)-path must therefore still be kept to compute the else-part also for t = W now. Moreover, we do not have CT(W -1) and we therefore assume min {CT(W - 1); 8*(W - 1) + C} = 8*(W - 1) + C (in eq. (9)). The explanation for this is as follows: A switch back in eq. (11) indicates that a new data distribution is established, such that the c-path that ends in a pdf state s from an old distribution routes its path through one of the more recent states that represent the new distribution, which means that this has lower costs despite of the incurred additional transition. Vice versa, a newly obtained pdf is unlikely to properly represent the previous mode then, which justifies our above assumption about CT (W -1). The effect of the proposed cut-off strategy is that we discard paths that end in pdfs from old modes but still allow to find the optimal pdf prototype within the current segment. Cut-off conditions occur shortly after mode changes in the data and cause the removal of HMM states with pdfs from old modes. However, if no mode change takes place in the incoming data sequence, no states will be discarded. We therefore still need to set a fixed upper limit", for the number of candidate paths/pdfs that are simultaneously under consideration if we only have limited resources available. When this limit is reached because no switches are detected, we must successively discard the oldest path/pdf stored, which finally might result in choosing a suboptimal prototype for that segment however. Ultimately, a continuous discarding even enforces a change of prototypes after 2", time steps if no switching is induced by the data until then. The buffer size", should therefore be as large as possible. In any case, the buffer overflow condition can be recorded along with the segmentation, which allows us to identify such artificial switchings. 2.5 The labeling algorithm A labeling algorithm is required to identify segments that represent the same underlying distribution and thus have similar pdf prototypes. The labeling algorithm generates labels for the segments and assigns identical labels to segments that are similar in this respect. To this end, we propose a relatively simple on-line clustering scheme for the prototypes, since we expect the prototypes obtained from the same underlying distribution to be already well-separated from the other prototypes as a result of the segmentation algorithm. We assign a new label to a segment if the distance of its associated prototype to all preceding prototypes exceeds a certain threshold and we assign the existing label of the closest preceding prototype otherwise. This can be written as e, l(R) = { ne.wlabel ,. if min1:'Sr<R {d(Pt(r) (x), Pt(R) (x))} > 1 (mdexmml:'Sr<R {d(Pt(r) (x), Pt(R) (x))} ), else; e (13) with the initialization l(l) = newlabel. Here, r = 1, ... , R, denotes the enumeration of the segments obtained so far , and t(?) denotes the mapping to the index of the corresponding pdf prototype. Note that the segmentation algorithm might replace a number of recent pdf prototypes (and also recent segmentation bounds) during the on-line processing in order to optimize the segmentation each time new data is presented. Therefore, a relabeling of all segments that have changed is necessary in each update step of the labeler. As for the hyperparameters (J and C, we developed an algorithm that computes a suitable value for from a sample set {X'd. We refer to our forthcoming publication [8]. 3 e Application We illustrate our approach by an application to a time series from switching dynamics based on the Mackey-Glass delay differential equation, dx(t) = -O.lx(t) dt 0.2x(t - td) . td)l? + 1 + x( t - (14) Eq. (14) describes a high-dimensional chaotic system that was originally introduced as a model of blood cell regulation [10]. In our example, four stationary operating modes, A, B, C, and D, are established by using different delays, td = 17, 23, 30, and 35, respectively. The dynamics operates stationary in one mode for a certain number of time steps, which is chosen at random between 200 and 300 (referring to sub-sampled data with a step size 6. = 6) . It then randomly switches to one of the other modes with uniform probability. This procedure is repeated 15 times, it thus generates a switching chaotic time series with 15 stationary segments. We then added a relatively large amount of "measurement" noise to the series: zero-mean Gaussian noise with a standard deviation of 30% of the standard deviation of the original series. The on-line segmentation algorithm was then applied to the noisy data, i.e. processing was performed on-line although the full data set was already available in this case. The scalar time series was embedded on-the-fly by using m = 6 and T = 1 (on the sub-sampled data) and we used a pdf window of size W = 50. The algorithm processed 457 data points per second on a 1.33 GHz PC in MATLAB/C under Linux, including the display of the ongoing segmentation, where one can observe the re-adaptation of past segmentation bounds and labels when new data becomes available. actual modes mode D modeC mode B mode A labels 1 2 3 4 3 561 3 3 6 2 2 bounds on-line segmentation xl!) Figure 1: Segmentation of a switching Mackey-Glass time series with noise (bottom) that operates in four different modes (top). The on-line segmentation algorithm (middle) , which receives the data points one by one, but not the mode information, yields correct segmentation bounds almost everywhere. The on-line labeler, however, assigns more labels (6) than desired (4) , presumably due to the fact that the segments are very short and noisy. The final segmentation is shown in Fig. 1. Surprisingly, the bounds of the segments are almost perfectly recovered from the very noisy data set. The only two exceptions are the third segment from the right , which is noticeably shorter than the original mode, and the segment in the middle, which is split in two by the algorithm. This split actually makes sense if one compares it with the data: there is a visible change in the signal characteristics at that point (t ~ 1500) even though the delay parameter was not modified there. This change is recorded by the algorithm since it operates in an unsupervised way. The on-line labeling algorithm correctly assigns single labels to modes A, B, and C, but it assigns three labels (4, 5, and 6) to the segments of mode D, the most chaotic one. This is probably due to the small sample sizes (of the segments), in combination with the large amount of noise in the data. 4 Discussion We presented an on-line method for the unsupervised segmentation and identification of sequential data, in particular from non-stationary switching dynamics. It is based on an HMM where the number of states varies dynamically as an effect of the way the incoming data is processed. In contrast to other approaches , it processes the data on-line and potentially even in real-time without training of any parameters. The method provides and updates a segmentation each time a new data point arrives. In effect, past segmentation bounds and labels are automatically re-adapted when new incoming data points are processed. The number of prototypes and labels that identify the segments is not fixed but determined by the algorithm. We expect useful applications of this method in fields where complex non-stationary dynamics plays an important role, like, e.g., in physiology (EEG, MEG), climatology, in industrial applications, or in finance. References [1] Bellman, R. E. (1957). Dynamic Programming, Princeton University Press, Princeton, NJ . [2] Bengio, Y, Frasconi, P. (1995). An Input Output HMM Architecture. In: Advances in Neural Information Processing Systems 7 (eds. Tesauro, Touretzky, Leen), Morgan Kaufmann, 427- 434. [3] Bengio, Y (1999). Markovian Models for Sequential Data. Neural Computing Surveys, http://www.icsi.berkeley.edu/~jagota/NCS, 2:129-162 . [4] Bishop, C. M. (1995). Neural Networks for Pattern Recognition , Oxford Univ. Press, NY. [5] Husmeier, D. (2000). Learning Non-Stationary Conditional Probability Distributions. Neural Networks 13, 287- 290. [6] Kehagias , A., Petridis, V. (1997). Time Series Segmentation using Predictive Modular Neural Networks. Neural Computation 9, 1691- 1710. [7] Kohlmorgen, J. , Miiller, K.-R., Rittweger, J. , Pawelzik, K. (2000). Identification of Nonstationary Dynamics in Physiological Recordings, Bioi Cybern 83(1),73- 84. [8] Kohlmorgen, J. , Lemm, S. , to appear. [9] Liehr, S., Pawelzik, K. , Kohlmorgen, J ., Miiller, K.-R. (1999). Hidden Markov Mixtures of Experts with an Application to EEG Recordings from Sleep. Theo Biosci 118, 246- 260. [10] Mackey, M., Glass, 1. (1977). Oscillation and Chaos in a Physiological Control System. Science 197, 287. [11] Packard, N. H., Crutchfield J. P. , Farmer, J . D. , Shaw, R. S. (1980). Geometry from a Time Series. Phys Rev Letters 45, 712- 716. [12] Pawelzik, K., Kohlmorgen, J. , Miiller, K.-R. (1996). Annealed Competition of Experts for a Segmentation and Classification of Switching Dynamics. Neural Computation 8(2), 340- 356. [13] Rabiner, L. R. (1989). A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition, Proceedings of the IEEE 77(2) , 257- 286. [14] Ramamurti, V., Ghosh, J. (1999). Structurally Adaptive Modular Networks for Non-Stationary Environments. IEEE Tr. Neural Networks 10(1), 152- 160.
2039 |@word middle:2 version:1 norm:1 termination:1 thereby:1 tr:1 initial:2 series:11 denoting:1 past:3 existing:1 current:1 ida:2 recovered:1 si:1 scatter:1 dx:2 must:7 readily:1 written:1 numerical:1 visible:1 remove:1 plot:1 update:10 v:2 stationary:8 mackey:3 selected:1 oldest:1 short:2 provides:1 lx:1 along:1 differential:1 introduce:1 market:1 bellman:1 td:3 resolve:1 cpu:3 enumeration:1 kohlmorgen:5 window:6 increasing:1 becomes:1 actual:1 automatically:1 underlying:5 notation:1 moreover:3 interpreted:1 developed:2 ghosh:1 nj:1 temporal:1 berkeley:1 y3:1 finance:1 control:1 farmer:1 appear:1 segmenting:1 before:1 limit:3 aiming:1 switching:12 despite:1 oxford:1 path:19 might:9 initialization:3 dynamically:2 limited:4 averaged:1 enforces:1 block:2 chaotic:3 procedure:1 area:1 physiology:1 projection:1 pre:1 get:2 cybern:1 www:2 optimize:1 demonstrated:1 yt:5 annealed:2 straightforward:1 survey:1 abrupt:1 assigns:4 estimator:1 financial:2 dw:1 embedding:6 pt:10 suppose:1 play:1 programming:2 us:1 recognition:2 particularly:1 updating:1 cut:3 bottom:1 steven:1 min1:1 fly:4 role:1 icsi:1 environment:1 dynamic:9 ultimately:1 depend:1 segment:20 predictive:1 logging:1 represented:1 separated:1 univ:1 detected:1 artificial:1 labeling:5 choosing:2 ise:1 modular:2 valued:2 s:3 reconstruct:1 otherwise:1 noisy:3 final:1 sequence:17 propose:4 product:1 adaptation:2 supposed:2 competition:2 p:2 optimum:1 incremental:2 help:1 illustrate:1 measured:2 nearest:1 eq:11 c:13 quantify:1 correct:1 centered:1 noticeably:1 sand:1 assign:2 probable:1 correction:1 considered:1 exp:2 presumably:1 viterbi:2 predict:1 mapping:1 consecutive:1 purpose:1 label:12 currently:1 vice:1 gaussian:2 aim:2 modified:1 avoid:1 varying:1 publication:2 improvement:1 pdfs:16 properly:1 likelihood:1 indicates:1 industrial:2 contrast:2 sense:1 glass:3 integrated:1 unlikely:1 hidden:3 germany:2 issue:1 classification:2 priori:2 smoothing:1 special:1 constrained:1 equal:1 once:3 construct:1 field:1 frasconi:1 sampling:2 labeler:2 identical:1 represents:1 unsupervised:4 simplify:1 inherent:1 ppt:1 randomly:1 simultaneously:3 relabeling:1 subsampled:1 phase:1 geometry:1 analyzed:1 mixture:2 arrives:2 pc:2 behind:1 necessary:2 respective:1 shorter:1 old:6 re:2 desired:1 minimal:2 classify:1 column:1 markovian:1 cost:11 deviation:2 uniform:2 usefulness:1 delay:4 too:1 reported:1 stored:1 varies:1 referring:1 density:5 yl:1 off:5 continuously:2 linux:2 squared:2 recorded:2 successively:1 choose:1 henceforth:1 expert:3 jek:2 potential:1 de:3 subsumes:1 kehagias:1 stream:6 performed:3 observing:1 reached:1 kekulestr:2 kaufmann:1 characteristic:1 efficiently:1 yield:2 identify:3 rabiner:1 identification:2 therewith:1 history:1 phys:1 touretzky:1 ed:1 definition:1 associated:3 sampled:5 newly:1 segmentation:32 actually:1 back:2 higher:2 dt:4 originally:1 leen:1 done:3 though:1 just:1 stage:1 until:1 d:4 hand:1 receives:1 nonlinear:1 incrementally:2 mode:19 effect:3 y2:1 analytically:1 regularization:1 during:3 width:1 climatological:1 pdf:17 complete:2 climatology:1 performs:1 chaos:1 novel:1 consideration:1 clause:1 refer:2 measurement:1 biosci:1 versa:1 automatic:1 tuning:1 unconstrained:1 similarly:1 similarity:1 operating:2 multivariate:1 closest:1 recent:5 optimizes:1 discard:3 scenario:1 store:1 buffer:2 certain:2 route:1 tesauro:1 jagota:1 jens:1 exploited:1 morgan:1 minimum:2 additional:1 preceding:2 husmeier:1 signal:2 sliding:2 multiple:1 full:2 exceeds:1 long:2 lin:1 equally:1 variant:2 iteration:2 kernel:2 represent:3 cell:1 addition:2 want:1 fine:1 completes:1 else:3 unlike:1 sr:2 probably:1 recording:3 hz:2 subject:1 induced:1 call:2 nonstationary:1 split:2 enough:3 bengio:2 switch:6 forthcoming:2 architecture:2 ifi:2 restrict:1 suboptimal:1 reduce:1 idea:3 prototype:18 perfectly:1 crutchfield:1 passed:1 reuse:1 effort:1 f:1 miiller:3 speech:2 cause:1 repeatedly:1 matlab:2 ignored:1 useful:2 involve:1 amount:6 processed:5 http:2 generate:1 lsi:1 tutorial:1 estimated:2 delta:1 track:3 per:1 correctly:1 hyperparameter:1 four:2 threshold:1 monitor:1 achieving:1 blood:1 changing:1 kept:1 merely:1 sum:1 everywhere:1 letter:1 logged:1 place:1 almost:2 oscillation:1 bound:9 ct:8 display:2 sleep:1 adapted:1 occur:2 constraint:1 kronecker:1 lemm:3 generates:2 speed:1 min:7 relatively:2 according:4 combination:1 smaller:1 describes:1 rev:1 b:2 restricted:1 resource:4 equation:2 previously:1 pawelzik:3 discus:1 turn:1 end:4 available:3 gaussians:1 operation:1 observe:1 appropriate:1 shaw:1 shortly:1 original:2 denotes:2 clustering:1 subsampling:1 top:1 neglect:2 exploit:1 overflow:1 already:4 added:1 fa:1 strategy:1 exhibit:3 distance:3 cw:1 berlin:2 hmm:18 originate:1 reason:1 induction:1 meg:2 length:1 index:3 modec:1 nc:1 regulation:1 potentially:2 perform:1 upper:2 observation:1 markov:3 discarded:1 tilde:1 rn:1 introduced:2 required:1 established:2 able:2 dynamical:4 pattern:1 including:2 memory:3 explanation:1 packard:1 suitable:1 scheme:1 brief:1 ne:1 fraunhofer:5 review:1 l2:1 removal:1 embedded:5 expect:3 filtering:1 proportional:1 incurred:1 sufficient:2 pij:1 storing:1 ift:1 changed:1 surprisingly:1 last:1 keeping:1 free:2 theo:1 allow:1 neighbor:1 ghz:2 dimension:2 calculated:1 world:1 transition:5 computes:2 perpetually:1 adaptive:1 far:3 sj:1 keep:2 incoming:8 don:1 continuous:2 iterative:1 rittweger:1 eeg:4 complex:1 main:2 noise:4 hyperparameters:1 repeated:1 fig:1 representative:1 fashion:1 ny:1 predictability:1 sub:3 structurally:1 xl:1 candidate:1 third:1 embed:1 specific:1 xt:7 discarding:2 bishop:1 physiological:3 consist:1 sequential:5 te:3 justifies:1 simply:2 likely:3 contained:1 scalar:1 determines:1 conditional:1 bioi:1 formulated:1 replace:2 change:9 hard:1 determined:2 operates:3 called:3 exception:1 internal:1 ongoing:1 princeton:2
1,140
204
308 Donnett and Smithers Neuronal Group Selection Theory: A Grounding in Robotics Jim Donnett and Tim Smithers Department of Artificial Intelligence University of Edinburgh 5 Forrest Hill Edinburgh EH12QL SCOTLAND ABSTRACT In this paper, we discuss a current attempt at applying the organizational principle Edelman calls Neuronal Group Selection to the control of a real, two-link robotic manipulator. We begin by motivating the need for an alternative to the position-control paradigm of classical robotics, and suggest that a possible avenue is to look at the primitive animal limb 'neurologically ballistic' control mode. We have been considering a selectionist approach to coordinating a simple perception-action task. 1 MOTIVATION The majority of industrial robots in the world are mechanical manipUlators - often arm-like devices consisting of some number of rigid links with actuators mounted where the links join that move adjacent links relative to each other, rotationally or translation ally. At the joints there are typically also sensors measuring the relative position of adjacent links, and it is in terms of position that manipulators are generally controlled (a desired motion is specified as a desired position of the end effector, from which can be derived the necessary positions of the links comprising the manipulator). Position control dominates largely for historical reasons, rooted in bang-bang control: manipulators bumped between mechanical stops placed so as to enforce a desired trajectory for the end effector. Neuronal Group Selection Theory: A Grounding in Robotics 1.1 SERVOMECHANISMS Mechanical stops have been superceded by position-controlling servomechanisms, negative feedback systems in which, for a typical manipulator with revolute joints, a desired joint angle is compared with a feedback signal from the joint sensor signalling actual measured angle; the difference controls the motive power output of the joint actuator proportionally. Where a manipulator is constructed of a number of links, there might be a servomechanism for each joint. In combination, it is well known that joint motions can affect each other adversely, requiring careful design and analysis to reduce the possibility of unpleasant dynamical instabilities. This is especially important when the manipulator will be required to execute fast movements involving many or all of the joints. We are interested in such dynamic tasks, and acknowledge some successful servomechanistic solutions (see [Andersson 19881, who describes a ping pong playing robot), but seek an alternative that is not as computationally expensive. 1.2 ESCAPING POSITION CONTROL In Nature, fast reaching and striking is a primitive and fundamental mode of control. In fast, time-optimal, neurologically ballistic movements (such as horizontal rotations of the head where subjects are instructed to turn it as fast as possible, [Hannaford and Stark 1985]), muscle activity patterns seem to show three phases: a launching phase (a burst of agonist), a braking phase (an antagonist burst), and a locking phase (a second agonist burst). Experiments have shown (see [Wadman et al. 1979]) that at least the first 100 mS of activity is the same even if a movement is blocked mechanically (without forewarning the subject), suggesting that the launch is specified from predetermined initial conditions (and is not immediately modified from proprioceptive information). With the braking and locking phases acting as a damping device at the end of the motion, the complete motion of the arm is essentially specified by the initial conditions - a mode radically differing from traditional robot positional control. The overall coordination of movements might even seem naive and simple when compared with the intricacies of servomechanisms (see [Braitenberg 1989, N ahvi and Hashemi 19841 who discuss the crane driver's strategy for shifting loads quickly and time-optimally). The concept of letting insights (such as these) that can be gained from the biological sciences shape the engineering principles used to create artificial autonomous systems is finding favour with a growing number of researchers in robotics. As it is not generally trivial to see how life's devices can be mapped onto machines, there is a need for some fundamental experimental work to develop and test the basic theoretical and empirical components of this approach, and we have been considering various robotics problems from this perspective. Here, we discuss an experimental two-link manipulator that performs a simple manipulation task - hitting a simple object perceived to be within its reach. The perception of the object specifies the initial conditions that determine an arm mo- 309 310 Donnett and Smithers tion that reaches it. In relating initial conditions with motor currents, we have been considering a scheme based on Neuronal Group Selection Theory [Edelman 1987, Reeke and Edelman 1988], a theory of brain organization. We believe this to be the first attempt to apply selectionist ideas in a real machine, rather than just in simulation. 2 NEURONAL GROUP SELECTION THEORY Edelman proposes Neuronal Group Selection (NGS) [Edelman 1978] as an organizing principle for higher brain function - mainly a biological basis for perception primarily applicable to the mammalian (and specifically, human) nervous system [Edelman 1981]. The essential idea is that groups of cells, structurally varied as a result of developmental processes, comprise a population from which are selected those groups whose function leads to adaptive behaviour of the system. Similar notions appear in immunology and, of course, evolutionary theory, although the effects of neuronal group selection are manifest in the lifetime of the organism. There are two premises on which the principle rests. The first is that the unit of selection is a cell group of perhaps 50 to 10,000 neurons. Intra-group connections between cells are assumed to vary (greatly) between groups, but other connections in the brain (particularly inter-group) are quite specific. The second premise is that the kinds of nervous systems whose organization the principle addresses are able to adapt to circumstances not previously encountered by the organism or its species [Edelman 1978]. 2.1 THREE CENTRAL TENETS There are three important ideas in the NGS theory [Edelman 1987]. ? A first selective process (cell division, migration, differentiation, or death) results in structural diversity providing a primary repertoire of variant cell groups. ? A second selective process occurs as the organism experiences its environment; group activity that correlates with adaptive behaviour leads to differential amplification of intra- and inter-group synaptic strengths (the connectivity pattern remains unchanged). From the primary repertoire are thus selected groups whose adaptive functioning means they are more likely to find future use - these groups form the ,econdary repertoire. ? Secondary repertoires themselves form populations, and the NGS theory additionally requires a notion of reentry, or connections between repertoires, usually arranged in maps, of which the well-known retinotopic mapping of the visual system is typical. These connections are critical for they correlate motor and sensory repertoires, and lend the world the kind of spatiotemporal continuity we all experience. Neuronal Group Selection Theory: A Grounding in Robotics 2.2 REQUffiEMENTS OF SELECTIVE SYSTEMS To be selective, a system must satisfy three requirements IReeke and Edelman 1988]. Given a configuration of input signals (ultimately from the sensory epithelia, but for 'deeper' repertoires mainly coming from other neuronal groups), if a group responds in a specific way it has matched the input IEdelman 1978]. The first requirement of a selective system is that it have a sufficiently large repertoire of variant elements to ensure that an adequate match can be found for a wide range of inputs. Secondly, enough of the groups in a repertoire must 'see' the diverse input signals effectively and quickly so that selection can operate on these groups. And finally, there must be a means for 'amplifying' the contribution, to the repertoire, of groups whose operation when matching input signals has led to adaptive behaviour. In determining the necessary number of groups in a repertoire, one must consider the relationship between repertoire size and the specificity of member groups. On the one hand, if groups are very specific, repertoires will need to be very large in order to recognize a wide range of possible inputs. On the other hand, if groups are not as discriminating, it will be possible to have smaller numbers of them, but in the limit (a single group with virtually no specificity) different signals will no longer be distinguishable. A simple way to quantify this might be to assume that each of N groups has a fixed probability, P, of matching an input configuration; then a typical measure IEdelman 1978] relating the effectiveness of recognition, r, to the number of groups is r = 1 - (1 - p)N (see Fig. 1). r log N Figure 1: Recognition as a Function of Repertoire Size From the shape of the curve in Fig. 1, it is clear that, for such a measure, below some lower threshold for N, the efficacy of recognition is equally poor. Similarly, above an upper threshold for N, recognition does not improve substantially as more groups are added. 3 SELECTIONISM IN OUR EXPERIMENT Our manipulator is required to touch an object perceived to be within reach. This is a well-defined but non-trivial problem in motor-sensory coordination. Churchland proposes a geometrical solution for his two-eyed 'crab' IChurchland 1986]' in which 311 312 Donnett and Smithers eye angles are mapped to those joint angles (the crab has a two-link arm) that would bring the end of the arm to the point currently foveated by the eyes. Such a novel solution, in which computation is implicit and massively parallel, would be welcome; however, the crab is a simulation, and no heed is paid to the question of how the appropriate sensory-motor mapping could be generated for a real arm. Reeke and Edelman discuss an automaton, Darwin III, similar to the crab, but which by selectional processes develops the ability to manipulate objects presented to it in its environment [Reeke and Edelman 19881. The Darwin III simulation does not account for arm dynamics; however, Edelman suggests that the training paradigm is able to handle dynamic effects as well as the geometry of the problem [Edelman 19891. We are attempting to implement a mechanical analogue of Darwin III, somewhat simplified, but which will experience the real dynamics of motion. S.l EXPERIMENTAL ARCHITECTURE AND HARDWARE The mechanical arrangement of our manipulator is shown in Fig. 2. The two links have agonist/antagonist tendon-drive arrangement, with an actuator per tendon. There are strain gauges in-line with the tendons. A manipulator 'reach' is specified by six parameters: burst amplitude and period for each of the three phases, launch, brake, and lock. 'I. 'I. 'I. 'I. ',tendons 'I. l upper-arm left actuator ,, , 'I. " '0 " Dri forearm/ left actuator \ U ~ upper-arm right actuator forearm right actuator Figure 2: Manipulator Mechanical Configuration Neuronal Group Selection Theory: A Grounding in Robotics At the end of the manipulator is an array of eleven pyroelectic-effect infrared detectors arranged in a U-shaped pattern. The relative location of a warm object presented to the arm is registered by the sensors, and is converted to eleven 8-bit integers. Since the sensor output is proportional to detected infrared energy flux, objects at the same temperature will give a more positive reading if they are close to the sensors than if they are further away. Also, a near object will register on adjacent sensors, not just on the one oriented towards it. Therefore, for a single, small object, a histogram of the eleven values will have a peak, and showing two things (Fig. 3): the sensor 'seeing' the most flux indicates the relative direction of the object, and the sharpness of the peak is proportional to the distance of the object. (object distant and to the left) (object near and straight ahead) Figure 3: Histograms for Distant Versus Near Objects Modelled on Darwin III [Reeke and Edelman 1988], the architecture of the selectional perception-action coordinator is as in Fig. 4. The boxes represent repertoires of appropriately interconnected groups of 'neurons'. Darwin III responds mainly to contour in a two-dimensional world, analogous to the recognition of histogram shape in our system. Where Darwin Ill's 'unique response' network is sensitive to line segment lengths and orientations, ours is sensitive to the length of subsequences in the array of sensor output values in which values increase or decrease by the same amount, and the amounts by which they change; similarly, where Darwin Ill's 'generic response' network is sensitive to presence of or changes in orientation of lines, ours responds to the presence of the subsequences mentioned above, and the positions in the array where two subsequences abut. The recognition repertoires are reciprocally connected, and both connect to the motor repertoire which consists of ballistic-movement 6-tuples. The system considers 'touching perceived object' to be adaptive, so when recognition activity correlates with a given 6-tuple, amplification ensures that the same response will be favoured in future. 313 314 Donnett and Smithers 4 WORK TO DATE As the sensing system is not yet functional, this aspect of the system is currently simulated in an IBM PC/AT. The rest of the electrical and mechanical hardware is in place. The major difficulty currently faced is that the selectional system will become computationally intensive on a serial machine. WORLD FEATURE DETECTOR FEATURE CORRELATOR classification couple COMBINATION RESPONSES (UNIQU~ ~ COMBINATION RESPONSES cim"':r"~f,~./(GENERIC) motor map MOTOR ACTIONS Figure 4: Experimental Architecture For each possible ballistic 'reach', there must be a representation for the 'reach 6-tuple'. Therefore, the motor repertoire must become large as the dexterity of the manipulator is increased. Similarly, as the array of sensors is extended (resolution increased, or field of view widened), the classification repertoires must also grow. On a serial machine, polling the groups in the repertoires must be done one at a time, introducing a substantial delay between the registration of object and the actual touch, precluding the interception by the manipulator of fast moving objects. We are exploring possibilities for parallelizing the selectional process (and have for this reason constructed a network of processing elements), with the expectation that this will lead us closer to fast, dynamic manipulation, at minimal computational expense. Neuronal Group Selection Theory: A Grounding in Robotics References Russell L. Andersson. A Robot Ping-Pong Player: Experiment in Real- Time Intelligent Control. MIT Press, Cambridge, MA, 1988. Valentino Braitenberg. "Some types of movement" , in C.G. Langton, ed., Artificial Life, pp. 555-565, Addison-Wesley, 1989. Paul M. Churchland. "Some reductive strategies in cognitive neurobiology". Mind, 95:279-309, 1986. Jim Donnett and Tim Smithers. "Behaviour-based control of a two-link ballistic arm". Dept. of Artificial Intelligence, University of Edinburgh, Research Paper RP .158, 1990. Gerald M. Edelman. "Group selection and phasic reentrant signalling: a theory of higher brain function", in G.M. Edelman and V.B. Mountcastle, eds., The Mindful Brain, pp. 51-100, MIT Press, Cambridge, MA, 1978. Gerald M. Edelman. "Group selection as the basis for higher brain function", in F.O. Schmitt et al., eds., Organization of the Cerebral Cortex, pp. 535-563, MIT Press, Cambridge, MA, 1981. Gerald M. Edelman. Neural Darwinism: The Theory of Neuronal Group Selection. Basic Books, New York, 1987. Gerald M. Edelman. Personal correspondence, 1989. Blake Hannaford and Lawrence Stark. "Roles of the elements of the triphasic control signal". Experimental Neurology, 90:619-634, 1985. M.J. Nahvi and M.R. Hashemi. "A synthetic motor control system; possible parallels with transformations in cerebellar cortex", in J .R. Bloedel et al., eds., Cerebellar Functions, pp. 67-69, Springer-Verlag, 1984. George N. Reeke Jr. and Gerald M. Edelman. "Real brains and artificial intelligence", in Stephen R. Graubard, ed., The Artificial Intelligence Debate, pp. 143-173, The MIT Press, Cambridge, MA, 1988. W.J. Wadman, J.J. Denier van der Gon, R.H. Geuse, and C.R. Mol. "Control of fast goal-directed arm movements". Journal of Human Movement Studies, 5:3-17, 1979. 315
204 |@word simulation:3 seek:1 paid:1 reentrant:1 initial:4 configuration:3 efficacy:1 ours:2 precluding:1 current:2 yet:1 must:8 tenet:1 distant:2 predetermined:1 eleven:3 shape:3 motor:9 intelligence:4 selected:2 device:3 nervous:2 signalling:2 scotland:1 location:1 launching:1 burst:4 constructed:2 differential:1 driver:1 become:2 edelman:20 consists:1 epithelium:1 inter:2 themselves:1 growing:1 brain:7 reductive:1 actual:2 correlator:1 considering:3 begin:1 retinotopic:1 matched:1 kind:2 substantially:1 differing:1 finding:1 transformation:1 differentiation:1 interception:1 control:14 unit:1 appear:1 positive:1 engineering:1 limit:1 might:3 suggests:1 range:2 directed:1 unique:1 implement:1 empirical:1 eyed:1 matching:2 specificity:2 seeing:1 suggest:1 onto:1 close:1 selection:15 applying:1 instability:1 crane:1 map:2 brake:1 primitive:2 automaton:1 sharpness:1 resolution:1 immediately:1 insight:1 array:4 valentino:1 his:1 population:2 handle:1 notion:2 autonomous:1 analogous:1 controlling:1 element:3 expensive:1 particularly:1 recognition:7 mammalian:1 infrared:2 gon:1 role:1 electrical:1 ensures:1 connected:1 movement:8 decrease:1 russell:1 mentioned:1 substantial:1 developmental:1 pong:2 locking:2 environment:2 dynamic:5 gerald:5 ultimately:1 personal:1 segment:1 churchland:2 division:1 basis:2 joint:9 various:1 fast:7 artificial:6 detected:1 whose:4 quite:1 ability:1 interconnected:1 coming:1 date:1 organizing:1 amplification:2 requirement:2 object:16 tim:2 develop:1 measured:1 launch:2 quantify:1 direction:1 bumped:1 human:2 premise:2 behaviour:4 repertoire:20 biological:2 secondly:1 exploring:1 sufficiently:1 crab:4 blake:1 lawrence:1 mapping:2 mo:1 major:1 vary:1 perceived:3 applicable:1 ballistic:5 currently:3 coordination:2 amplifying:1 sensitive:3 create:1 gauge:1 mit:4 sensor:9 modified:1 reaching:1 rather:1 derived:1 indicates:1 mainly:3 greatly:1 industrial:1 rigid:1 typically:1 coordinator:1 selective:5 comprising:1 interested:1 polling:1 overall:1 motive:1 ill:2 orientation:2 classification:2 proposes:2 animal:1 field:1 comprise:1 shaped:1 ng:3 look:1 future:2 braitenberg:2 develops:1 intelligent:1 primarily:1 oriented:1 recognize:1 phase:6 consisting:1 geometry:1 attempt:2 organization:3 possibility:2 intra:2 pc:1 tuple:2 closer:1 necessary:2 experience:3 damping:1 desired:4 theoretical:1 darwin:7 minimal:1 effector:2 increased:2 measuring:1 organizational:1 introducing:1 abut:1 delay:1 successful:1 motivating:1 optimally:1 connect:1 spatiotemporal:1 synthetic:1 migration:1 fundamental:2 immunology:1 discriminating:1 peak:2 mechanically:1 quickly:2 connectivity:1 central:1 langton:1 adversely:1 cognitive:1 book:1 stark:2 suggesting:1 account:1 converted:1 diversity:1 satisfy:1 register:1 tion:1 view:1 parallel:2 contribution:1 largely:1 who:2 modelled:1 agonist:3 trajectory:1 researcher:1 drive:1 straight:1 ping:2 donnett:6 reach:6 detector:2 synaptic:1 ed:5 energy:1 pp:5 couple:1 stop:2 manifest:1 amplitude:1 wesley:1 higher:3 response:5 arranged:2 execute:1 box:1 done:1 lifetime:1 just:2 implicit:1 hand:2 ally:1 horizontal:1 touch:2 continuity:1 mode:3 perhaps:1 believe:1 manipulator:16 grounding:5 effect:3 requiring:1 concept:1 functioning:1 death:1 proprioceptive:1 adjacent:3 rooted:1 m:1 antagonist:2 hill:1 complete:1 performs:1 motion:5 bring:1 temperature:1 geometrical:1 dexterity:1 novel:1 rotation:1 reeke:5 heed:1 functional:1 neurologically:2 cerebral:1 organism:3 relating:2 braking:2 blocked:1 cambridge:4 similarly:3 mindful:1 moving:1 robot:4 longer:1 cortex:2 touching:1 perspective:1 manipulation:2 massively:1 verlag:1 graubard:1 life:2 der:1 muscle:1 rotationally:1 george:1 somewhat:1 determine:1 paradigm:2 period:1 signal:6 stephen:1 match:1 adapt:1 serial:2 equally:1 manipulate:1 controlled:1 involving:1 basic:2 variant:2 essentially:1 circumstance:1 smithers:6 expectation:1 histogram:3 represent:1 cerebellar:2 robotics:8 cell:5 grow:1 appropriately:1 rest:2 operate:1 subject:2 virtually:1 dri:1 thing:1 member:1 seem:2 effectiveness:1 call:1 integer:1 structural:1 near:3 presence:2 hashemi:2 iii:5 enough:1 affect:1 architecture:3 escaping:1 reduce:1 idea:3 avenue:1 intensive:1 favour:1 six:1 york:1 action:3 adequate:1 generally:2 proportionally:1 clear:1 selectional:4 amount:2 servomechanism:4 welcome:1 hardware:2 specifies:1 coordinating:1 per:1 diverse:1 group:39 threshold:2 registration:1 angle:4 striking:1 place:1 forrest:1 tendon:4 bit:1 correspondence:1 encountered:1 activity:4 strength:1 ahead:1 aspect:1 attempting:1 department:1 combination:3 poor:1 jr:1 describes:1 smaller:1 computationally:2 previously:1 remains:1 discus:4 turn:1 phasic:1 addison:1 letting:1 mind:1 end:5 operation:1 apply:1 limb:1 actuator:7 away:1 enforce:1 appropriate:1 generic:2 alternative:2 rp:1 ensure:1 lock:1 especially:1 classical:1 unchanged:1 move:1 added:1 question:1 occurs:1 arrangement:2 strategy:2 primary:2 traditional:1 responds:3 evolutionary:1 distance:1 link:11 mapped:2 simulated:1 majority:1 considers:1 trivial:2 reason:2 length:2 relationship:1 providing:1 expense:1 debate:1 negative:1 design:1 upper:3 neuron:2 forearm:2 acknowledge:1 extended:1 neurobiology:1 head:1 jim:2 strain:1 varied:1 parallelizing:1 mechanical:7 specified:4 required:2 connection:4 widened:1 registered:1 address:1 able:2 dynamical:1 perception:4 pattern:3 usually:1 below:1 reading:1 lend:1 reciprocally:1 shifting:1 power:1 critical:1 analogue:1 difficulty:1 warm:1 arm:12 scheme:1 improve:1 eye:2 cim:1 naive:1 denier:1 faced:1 mountcastle:1 determining:1 relative:4 revolute:1 mounted:1 proportional:2 versus:1 schmitt:1 principle:5 playing:1 translation:1 ibm:1 course:1 placed:1 deeper:1 wide:2 edinburgh:3 van:1 feedback:2 curve:1 world:4 contour:1 sensory:4 instructed:1 adaptive:5 simplified:1 historical:1 flux:2 correlate:3 robotic:1 assumed:1 tuples:1 neurology:1 subsequence:3 additionally:1 nature:1 mol:1 motivation:1 paul:1 neuronal:12 fig:5 join:1 favoured:1 structurally:1 position:9 load:1 specific:3 showing:1 sensing:1 dominates:1 essential:1 effectively:1 gained:1 foveated:1 intricacy:1 led:1 distinguishable:1 likely:1 positional:1 visual:1 hitting:1 springer:1 radically:1 ma:4 goal:1 bang:2 careful:1 towards:1 change:2 typical:3 specifically:1 acting:1 andersson:2 specie:1 secondary:1 experimental:5 player:1 unpleasant:1 dept:1
1,141
2,040
PAC Generalization Bounds for Co-training Sanjoy Dasgupta AT&T Labs?Research dasgupta@research.att.com Michael L. Littman AT&T Labs?Research mlittman@research.att.com David McAllester AT&T Labs?Research dmac@research.att.com Abstract The rule-based bootstrapping introduced by Yarowsky, and its cotraining variant by Blum and Mitchell, have met with considerable empirical success. Earlier work on the theory of co-training has been only loosely related to empirically useful co-training algorithms. Here we give a new PAC-style bound on generalization error which justifies both the use of confidences ? partial rules and partial labeling of the unlabeled data ? and the use of an agreement-based objective function as suggested by Collins and Singer. Our bounds apply to the multiclass case, i.e., where instances are to be assigned one of labels for .  1 Introduction In this paper, we study bootstrapping algorithms for learning from unlabeled data. The general idea in bootstrapping is to use some initial labeled data to build a (possibly partial) predictive labeling procedure; then use the labeling procedure to label more data; then use the newly labeled data to build a new predictive procedure and so on. This process can be iterated until a fixed point is reached or some other stopping criterion is met. Here we give PAC style bounds on generalization error which can be used to formally justify certain boostrapping algorithms. One well-known form of bootstrapping is the EM algorithm (Dempster, Laird and Rubin, 1977). This algorithm iteratively updates model parameters by using the current model to infer (a probability distribution on) labels for the unlabeled data and then adjusting the model parameters to fit the (distribution on) filled-in labels. When the model defines a joint probability distribution over observable data and unobservable labels, each iteration of the EM algorithm can be shown to increase the probability of the observable data given the model parameters. However, EM is often subject to local minima ? situations in which the filled-in data and the model parameters fit each other well but the model parameters are far from their maximum-likelihood values. Furthermore, even if EM does find the globally optimal maximum likelihood parameters, a model with a large number of parameters will over-fit the data. No PAC-style guarantee has yet been given for the generalization accuracy of the maximum likelihood model. An alternative to EM is rule-based bootstrapping of the form used by Yarowsky (1995), in which one assigns labels to some fraction of a corpus of unlabeled data and then infers new labeling rules using these assigned labels as training data. New labels lead to new rules which in turn lead to new labels, and so on. Unlike EM, rule-based bootstrapping typically does not attempt to fill in, or assign a distribution over, labels unless there is compelling evidence for a particular label. One intuitive motivation for this is that by avoiding training on low-confidence filled-in labels one might avoid the self-justifying local optima encountered by EM. Here we prove PAC-style generalization guarantees for rulebased bootstrapping. Our results are based on an independence assumption introduced by Blum and Mitchell (1998) which is rather strong but is used by many successful applications. Consider, for example, a stochastic context-free grammar. If we generate a parse tree using such a grammar then the nonterminal symbol labeling a phrase separates the phrase from its context ? the phrase and the context are statistically independent given the nonterminal symbol. More intuitively, in natural language the distribution of contexts into which a given phrase can be inserted is determined to some extent by the ?type? of the phrase. The type includes the syntactic category but might also include semantic subclassifications, for instance, whether a noun phrase refers to a person, organization, or location. If we think of each particular occurrence of a phrase as a triple     , where  is the phrase itself,  is the ?type? of the phrase, and  is the context, then we expect that  is conditionally independent of  given  . The conditional independence can be made to hold precisely if we generate such triples using a stochastic context free grammar where  is the syntactic category of the phrase. Blum and Mitchell introduce co-training as a general term for rule-based bootstrapping in which each rule must be based entirely on   or entirely on  . In other words, there are two distinct hypothesis classes,   which consists of functions predicting  from   , and  which consists of functions predicting  from  . A co-training algorithm bootstraps by alternately selecting     and    . The principal assumption made by Blum and Mitchell is that   is conditionally independent of  given  . Under such circumstances, they show that, given a weak predictor in   , and given an algorithm which can learn  under random misclassification noise, it is possible to learn a good predictor in  . This gives some degree of justification for the co-training restriction on rule-based bootstrapping. However, it does not provide a bound on generalization error as a function of empirically measurable quantities. Furthermore, there is no apparent relationship between this PAC-learnability theorem and the iterative co-training algorithm they suggest. Collins and Singer (1999) suggest a refinement of the co-training algorithm in which one explicitly optimizes an objective function that measures the degree of agreement between the predictions based on   and those based on  . They describe methods for ?boosting? this objective function but do not provide any formal justification for the objective function itself. Here we give a PAC-style performance guarantee in terms of this agreement rate. This guarantee formally justifies the Collins and Singer suggestion. In this paper, we use partial classification rules, which either output a class label or output a special symbol  indicating no opinion. The error of a partial rule is the probability that the rule is incorrect given that it has an opinion. We work in the co-training setting where we have a pair of partial rules  and  where   (sometimes) predicts  from  and  (sometimes) predicts  from  . Each of the rules  and  can be ?composite rules?, such as decision lists, where each composite rule contains a large set of smaller rules within it. We give a bound on the generalization error of each of the rules   and  in terms of the empirical agreement rate between the two rules. This bound formally justifies both the use h1 X1 X2 h2 Y Figure 1: The co-training scenario with rules  and  . of agreement in the objective function and the use of partial rules. The bound shows the potential power of unlabeled data ? low generalization error can be achieved for complex rules with a sufficient quantity of unlabeled data. The use of partial rules is analogous to the use of confidence ratings ? a partial rule is just a rule with two levels of confidence. So the bound can also be viewed as justifying the partial labeling aspect of rule-based bootstrapping, at least in the case of co-training where an independence assumption holds. The generalization bound leads naturally to algorithms for optimizing the bound. A simple greedy procedure for doing this is quite similar to the co-training algorithm suggested by Collins and Singer. 2 The Main Result We start with some basic definitions and observations. Let be an i.i.d. sample consisting of individual samples  ,  ,  . For any statement    we let  be the subset        . For any two statements  and  we define the empirical estimate   to be    !  . For the co-training bounds proved here we assume data is drawn from some distribution over triples       with   #"     %$    , and  &" , and   where   and  are conditionally independent given  , that is,       ('       and       )'    . In the co-training framework we are given an unlabeled sample +* of pairs     drawn i.i.d. from the underlying distribution, and possibly some labeled samples +, . We will mainly be interested in making inferences from the unlabeled data. A partial rule  on a set " is a mapping from " to %$    - . We will be interested in pairs of partial rules  and  which largely agree on the unlabeled data. The conditional probability relationships in our scenario are depicted graphically in figure 1. Important intuition is given by the data-processing inequality of information theory .2  30   . In other words, any mutual information (Cover and Thomas, 1991): ./ 10  between   and  must be mediated through  . In particular, if  and  agree to a large extent, then they must reveal a lot about  . And yet finding such a pair      requires no labeled data at all. This simple observation is a major motivation for the proof, but things are complicated considerably by partial rules and by approximate agreement.  For a given partial rule   with   (' 4 6587 9: <;<='?>A@BDC E1F H GJIKGML  define a function 9 on %$       by 'ON P  '?;< We want   to be a nearly deterministic function of the actual label  ; in other words, we  want   'Q9:    ' 4  to be near one. We would also like  to carry the same information as  . This is equivalent to saying that 9 should be a permutation of the possible labels A$    . Here we give a condition using only unlabeled data which guarantees, up to high confidence, that 9 is a permutation; this is the best we can hope for using unlabeled   data alone. We also bound the error rates  D' 4 N :9:  ' N   ' 4  using only unlabeled data. In the case of ' , if 9 is a permutation then 9 is either the identity function or the function reversing the two possible values. We use the unlabeled data to select   and  so that 9 is a permutation and   has low error rate. We can then use a smaller amount of labeled data to determine which permutation we have found. We now introduce a few definitions related to sampling issues. Some measure of the complexity of rules  and  is needed; rather than VC dimension, we adopt a clean notion of bit length. We assume that rules are specified in some rule language and write    for the number of bits used to specify the rule  . We assume that the rule language is prefix-free (no proper prefix of the bit string specifying a rule is itself a legal rule specification). A $ . For given partial rules   and prefix free code satisfies the Kraft inequality  and N  A$    we now define the following functions of the sample . The first, as we will see, is a bound on the sampling error for empirical probabilities conditioned upon  'ON    ' 4  . The second is a sampling-adjusted disagreement rate between   and  .       !#"$ % %$&'% (% )&*! +    "-% .   0/1 352 467% , 8    9  10/%:;/ ) '2 46< 9   ) 02 /=%:0/> 52 43?<@"  A( $3C  Note that if the sample size is sufficiently large (relative to   and  % ) then HB I is near zero. Also note that if and have near perfect agreement when they both are $3C  is near one. We can now state our main result. not then D I Theorem 1 With probability at least $FE5C  over the choice of the sample , we have that for all  and, , if D I GC3 5 7 for $ N then (a) 9 is a permutation and (b) for N all $    -'?4 N6 ' N -' 4 IH5B I G3C    4 N639: ='ON -' 4  -'? D I G3C                                               The theorem states, in essence, that if the sample size is large, and   and  largely agree  on the unlabeled data, then   ' 4 N6  '?N   -' 4  is a good estimate of the error rate  4 N639:  ='ON   -' 4  .  -'? The theorem also justifies the use of partial rules. Of course it is possible to convert a partial rule to a total rule by forcing a random choice when the rule would otherwise return  . Converting a partial rule to a total rule in this way and then applying the above theorem to the total rule gives a weaker result. An interesting case is when ' ,  is total and is  a perfect copy of  , and  -' 4  happens to be $1!   . In this case the empirical error rate of the corresponding total rule ? the rule that guesses when  has no opinion ? will be statistically indistinguishable from from 1/2. However, in this case theorem 1 can still establish that the false positive and false negative rate of the partial rule  is near zero. J  3 The Analysis We start with a general lemma about conditional probability estimation. Lemma 2 For any i.i.d. sample , and any statements  and  about individual instances in the sample, the following holds with probability at least $ over the choice of . M MM   1E A    KELD  P N O  MM    !D    M (1)  MMM   A1E      MMM 5         M '  L    '   MM    1E         '  D ' D L Proof. We have the following where the third step follows by the Chernoff bound.  MM MM     M 5      >   KE    1  C ! ,  1E 'ON6 'ON MM Therefore, with probability at least $    '  BI GC3 (2) for any given . By the union bound and the Kraft inequality, we have that with  N  . probability at least $ E C this must hold simultaneously for all and , and all $ Lemma 3 Pick  any rules and for which equation (2) as well as D I G3C 5 7    '   N 'ON '4         -' 4                hold for all $ N        . Then 9 is a permutation, and moreover, for any N ,    -' 4 '?N 39:  =  '?N    5 $1!    Pick any N  %$    . We need to show that there exists some such that 9: ='?N . By the definition of I and condition (2) we know   4 N  ' N    ' 4    ? ' N  ' N    ' 4    '? I      3  Since I      3 587 , it follows that   '?N6  '?N    ' 4 5 $! . Rewriting this  D Proof. D  D E GC by conditioning on  , we get  G GML   P  '  'ON   -' 4   ! 'ON P  '     -' 4  5   $C   $!    " The summation is a convex combination; therefore there must exist some such that    'QN  '   )' 4  5 $! . So for each N there must exist a with 9: %Q ' N, whereby 9 is a permutation. Lemma 4 Pick any rules   <9:   ='ON  'ON     -' 4 D $C and  satisfying the conditions of the previous lemma. Then is at least I      3 .  Proof. By the previous lemma 9 is a permutation, so 9:  has the same information content as  . Therefore   and  are conditionally independent given 9:  . For any N , D I  GC3        '      '?N   ! ' <9:  =  ' <9:   ='ON  H $   # I  N  'ON     'ON    4 N6  -'?   -' 4   -' 4 'ON % 9: =' 1E   -' 4     '4 E       E    ' N   -' 4  ! ! 'ON A9:  ' N6 9: ='  -' 4   -' 4   -' 4     'ON639: ='?N   -' 4  N39: ='?N    ' 4     'ON639: ='    ' 4     ' 39: ='    ' 4   -' 4 E %     where the second step involves conditioning on 9:  . Also by the previous lemma, we  have   ' & 9: '    '4  5 $! so the second term in the above sum must be negative, whereby  D I  GC3          <9:  =  'ON6 '?N   -' 4  <9:  =  'ON6 '?N   -' 4    E  'ON  9: ='ON   4 N6 9: =' N '?     '4     -' 4   Under the conditions of these lemmas, we can derive the bounds on error rates:  4 N39: ='  -'O   -' 4 N     ' 4 N6  ? ' N  9: ='ON6 ' N     4 N6 ' '?   N D I '4    '4       '4  IH5BHI  GC3 GC3           4 Bounding Total Error Assuming that we make a random guess when written as follows.   ='    -' 4    -' 4 '    '4 9:   -  I GC3 D I '      $  GC3         4 N  'O    can be =H 8E $  '  G3C  to be the bounds on the  To give a bound on the total error rate we first define I error rate for label N given in theorem 1.  , the total error rate of         'ON    '4 IH'BHI  GC3      We can now state a bound on the total error rate as a corollary to theorem 1. E C GC  Corollary 5 With probability at least $ over the choice of we have the follow and  such that for all N we have I      !  5 7 and ing for all pairs of rules  I      !  8$ ! . GC     E     E B   GC !    C  E1F   H E8$   ' IH'B   $C !     NPO  N O  H ! C    (' 4      B $C3  D '      GC !          E C  $C  Proof. From our main theorem, we know that with probability at least $ ! , for all   N .  #' 4 N-M9:  ' N   #' 4  is bounded by I      !  . This implies that with probability at least $ ! ,    KE0C      '4  C 1 E F   A GC !  IH 8E $ $KE         '4   (3) GC  E C   E  With probability at least $ ! we have that     ' 4    ' 4  is no larger than       !  . So by the union bound both of these conditions hold simultaneously with  8$ ! we have that the upper probability at least $ . Since CPE3F% 3     !    bound in (3) is maximized by setting  -' 4  equal to   (' 4        !  . B KE0C $C   E E B GC  Corollary 5 can be improved in a variety of ways. One could use a relative Chernoff  bound to tighten the uncertainty in   ' 4  in the case where this probability is small.   One could also use the error rate bounds I      !  to construct bounds on 9: ='  N6  -' 4  . One could then replace the max over I      !  by a convex combination. Another approach is to use the error rate of a rule that combines   and  , e.g., the rule outputs   if  &' 4  , otherwise outputs  if  )' 4  , and otherwise guesses a random value. This combined rule will have a lower error rate and it is possible to give bounds on the error rate of the combined rule. We will not pursue these refined bounds here. It should be noted, however, that the algorithm described in section 4 can be used with any bound on total error rate. $C  $C  5 A Decision List Algorithm This section suggests a learning algorithm inspired by both Yarowsky (1995) and Collins and Singer (1999) but modified to take into account theorem 1 and Corollary 5. Corollary 5, or some more refined bound on total error, provides an objective function that can be pursued by a learning algorithm ? the objective is to find   and  so as to minimize the upper bound on the total error rate. Typically, however, the search space is enormous. Following Yarowsky, we consider the greedy construction of decision list rules.  and    "  we have      be two ?feature sets? such that for    and  )" we have     7 $3 . We assume that   is to be a decision list over the features in  , i.e., a finite sequence of the form    ;  ; /L  L  and 3I  A$    . A decision list can be viewed as a right-branching where /I  decision tree. More specifically, if  is the list     ;    ;  ; /L  L then  1   is   if 1  'Q$ and otherwise equals the value of the list   ;  ; JL  L on   . We define an empty decision list to have value  . For in <7 $ we can define    as follows where is the number of feature-value pairs in   . Let  and 7  $  and for  N BH $ N B  3 KE0B equals the probability that a  certain stochastic process      I $ which . This implies the Kraft inequality is all that is     '   1   It is possible to show that $ generates the rule   needed in theorem 1 and corollary 5. We also assume that features and define    similarly.  is a decision list over the Following Yarowsky we suggest growing the decision lists in a greedy manner adding one feature value pair at a time. A natural choice of greedy heuristic might be a bound on the total error rate. However, in many cases the final objective function is not an appropriate choice for the greedy heuristic in greedy algorithms. A* search, for example, might be viewed as a greedy heuristic where the heuristic function estimates the number of steps needed to reach a high-value configuration ? a low value configuration might be one step away from a high value configuration. The greedy heuristic used in greedy search should estimate the value of the final configuration. Here we suggest using CPE3F        !  as a heuristic estimate of the final total error rate ? in the final configuration we should GC    GC have that  -' 4  is reasonably large and the important term will be C E1F        3 . For concreteness, we propose the following algorithm. Many variants of this algorithm also seem sensible. 1. Initialize   and knowledge.  to ?seed rule? decision lists using domain-specific prior   2. Until    '  and   ' both rules, do the following.  GC  O   are both zero, or all features have been used in  (a) Let B denote   if    ' 65   '  and  otherwise. 7 for some N , then extend B by the pair  which (b) If I      !  I I      ! . most increases C (c) Otherwise extend B by a single feature-value pair selected to minimize the  C E1FA A      !  . D  GC  D $C  3. Prune the rules ? iteratively remove the pair from the end of either  or  that greedily minimizes the bound on total error until no such removal reduces the bound. 6 Future Directions We have given some theoretical justification for some aspects of co-training algorithms that have been shown to work well in practice. The co-training assumption we have used in our theorems are is at best only approximately true in practice. One direction for future research is to try to relax this assumption somehow. The co-training assumption states that   and  are independent given  . This is equivalent to the statement that the mutual information between  and  given  is zero. We could relax this assumption by allowing some small amount of mutual information between   and  given  and giving bounds on error rates that involve this quantity of mutual information. Another direction for future work, of course, is the empirical evaluation of co-training and bootstrapping methods suggested by our theory. Acknowledgments The authors wish to acknowledge Avrim Blum for useful discussions and give special thanks to Steve Abney for clarifying insights. Literature cited Blum, A. & Mitchell, T. (1998) Combining labeled and unlabeled data with co-training. COLT. Collins, M. & Singer, Y. (1999) Unsupervised models for named entity classification. EMNLP. Cover, T. & Thomas, J. (1991) Elements of information theory. Wiley. Dempster, A., Laird, N. & Rubin, D. (1977) Maximum-likelihood from incomplete data via the EM algorithm. J. Royal Statist. Soc. Ser. B, 39:1-38. Nigam, K. & Ghani, R. (2000) Analyzing the effectiveness and applicability of co-training. CIKM. Yarowsky, D. (1995) Unsupervised word sense disambiguation rivaling supervised methods. ACL.
2040 |@word pick:3 carry:1 initial:1 configuration:5 contains:1 att:3 selecting:1 prefix:3 current:1 com:3 yet:2 must:7 written:1 remove:1 update:1 alone:1 greedy:9 pursued:1 guess:3 selected:1 provides:1 boosting:1 location:1 incorrect:1 prove:1 consists:2 combine:1 manner:1 introduce:2 growing:1 inspired:1 globally:1 actual:1 underlying:1 moreover:1 bounded:1 string:1 pursue:1 minimizes:1 finding:1 bootstrapping:11 guarantee:5 ser:1 yarowsky:6 positive:1 local:2 analyzing:1 approximately:1 might:5 acl:1 on6:6 specifying:1 suggests:1 co:20 bi:1 statistically:2 acknowledgment:1 union:2 practice:2 bootstrap:1 procedure:4 empirical:6 composite:2 confidence:5 word:4 refers:1 suggest:4 get:1 unlabeled:15 bh:1 context:6 applying:1 restriction:1 measurable:1 deterministic:1 equivalent:2 graphically:1 convex:2 ke:2 assigns:1 rule:60 insight:1 fill:1 notion:1 justification:3 analogous:1 construction:1 hypothesis:1 agreement:7 element:1 satisfying:1 rivaling:1 predicts:2 labeled:6 inserted:1 e8:1 intuition:1 dempster:2 complexity:1 littman:1 predictive:2 upon:1 kraft:3 joint:1 distinct:1 describe:1 labeling:6 refined:2 apparent:1 quite:1 larger:1 heuristic:6 relax:2 otherwise:6 grammar:3 syntactic:2 think:1 itself:3 laird:2 final:4 a9:1 sequence:1 propose:1 combining:1 intuitive:1 empty:1 optimum:1 perfect:2 derive:1 mmm:2 nonterminal:2 strong:1 soc:1 involves:1 implies:2 met:2 direction:3 stochastic:3 vc:1 mcallester:1 opinion:3 assign:1 generalization:9 summation:1 adjusted:1 hold:6 mm:6 sufficiently:1 seed:1 mapping:1 major:1 adopt:1 estimation:1 label:16 hope:1 modified:1 rather:2 avoid:1 corollary:6 likelihood:4 mainly:1 greedily:1 sense:1 inference:1 stopping:1 typically:2 interested:2 unobservable:1 classification:2 issue:1 colt:1 noun:1 special:2 initialize:1 mutual:4 equal:3 construct:1 sampling:3 chernoff:2 unsupervised:2 nearly:1 future:3 few:1 simultaneously:2 individual:2 consisting:1 attempt:1 organization:1 evaluation:1 partial:19 unless:1 filled:3 tree:2 loosely:1 incomplete:1 theoretical:1 instance:3 earlier:1 compelling:1 cover:2 phrase:10 applicability:1 subset:1 predictor:2 successful:1 learnability:1 considerably:1 combined:2 person:1 thanks:1 cited:1 michael:1 possibly:2 emnlp:1 e1f:3 style:5 return:1 account:1 potential:1 m9:1 includes:1 explicitly:1 h1:1 lot:1 lab:3 try:1 doing:1 reached:1 start:2 complicated:1 minimize:2 accuracy:1 largely:2 gc3:9 maximized:1 weak:1 iterated:1 reach:1 definition:3 npo:1 naturally:1 proof:5 newly:1 proved:1 adjusting:1 mitchell:5 knowledge:1 infers:1 steve:1 supervised:1 follow:1 specify:1 improved:1 furthermore:2 just:1 until:3 parse:1 somehow:1 defines:1 reveal:1 true:1 assigned:2 iteratively:2 semantic:1 conditionally:4 indistinguishable:1 self:1 branching:1 essence:1 whereby:2 noted:1 criterion:1 empirically:2 conditioning:2 jl:1 extend:2 similarly:1 language:3 specification:1 mlittman:1 optimizing:1 optimizes:1 forcing:1 scenario:2 certain:2 inequality:4 success:1 minimum:1 converting:1 prune:1 determine:1 infer:1 reduces:1 ing:1 justifying:2 prediction:1 variant:2 basic:1 circumstance:1 iteration:1 sometimes:2 achieved:1 want:2 unlike:1 subject:1 thing:1 seem:1 effectiveness:1 near:5 hb:1 variety:1 independence:3 fit:3 idea:1 multiclass:1 whether:1 useful:2 involve:1 amount:2 statist:1 category:2 generate:2 exist:2 cikm:1 write:1 dasgupta:2 blum:6 enormous:1 drawn:2 clean:1 rewriting:1 concreteness:1 fraction:1 convert:1 sum:1 uncertainty:1 named:1 saying:1 disambiguation:1 decision:10 bit:3 entirely:2 bound:34 q9:1 encountered:1 precisely:1 rulebased:1 x2:1 generates:1 aspect:2 bdc:1 combination:2 smaller:2 em:8 making:1 happens:1 intuitively:1 legal:1 equation:1 agree:3 turn:1 singer:6 needed:3 know:2 end:1 apply:1 away:1 appropriate:1 disagreement:1 occurrence:1 alternative:1 thomas:2 include:1 giving:1 build:2 establish:1 objective:8 quantity:3 separate:1 clarifying:1 entity:1 sensible:1 extent:2 assuming:1 length:1 code:1 relationship:2 bhi:1 statement:4 negative:2 proper:1 allowing:1 upper:2 observation:2 finite:1 acknowledge:1 situation:1 gc:12 rating:1 david:1 introduced:2 pair:10 specified:1 c3:1 alternately:1 suggested:3 max:1 royal:1 power:1 misclassification:1 natural:2 predicting:2 mediated:1 n6:11 prior:1 literature:1 removal:1 relative:2 expect:1 permutation:9 suggestion:1 interesting:1 triple:3 h2:1 degree:2 sufficient:1 rubin:2 course:2 free:4 copy:1 formal:1 weaker:1 dimension:1 qn:1 author:1 made:2 refinement:1 far:1 tighten:1 approximate:1 observable:2 corpus:1 search:3 iterative:1 abney:1 learn:2 reasonably:1 nigam:1 complex:1 domain:1 main:3 motivation:2 noise:1 bounding:1 ghani:1 x1:1 wiley:1 wish:1 cotraining:1 third:1 theorem:12 specific:1 pac:7 symbol:3 list:11 evidence:1 exists:1 ih:3 false:2 adding:1 avrim:1 justifies:4 conditioned:1 depicted:1 gml:1 satisfies:1 conditional:3 viewed:3 identity:1 replace:1 considerable:1 content:1 determined:1 specifically:1 reversing:1 justify:1 principal:1 lemma:8 total:15 sanjoy:1 indicating:1 formally:3 select:1 collins:6 avoiding:1
1,142
2,041
Grammatical Bigrams Mark A. Paskin Computer Science Division University of California, Berkeley Berkeley, CA 94720 paskin@cs.berkeley.edu Abstract Unsupervised learning algorithms have been derived for several statistical models of English grammar, but their computational complexity makes applying them to large data sets intractable. This paper presents a probabilistic model of English grammar that is much simpler than conventional models, but which admits an efficient EM training algorithm. The model is based upon grammatical bigrams, i.e. , syntactic relationships between pairs of words. We present the results of experiments that quantify the representational adequacy of the grammatical bigram model, its ability to generalize from labelled data, and its ability to induce syntactic structure from large amounts of raw text. 1 Introduction One of the most significant challenges in learning grammars from raw text is keeping the computational complexity manageable. For example, the EM algorithm for the unsupervised training of Probabilistic Context-Free Grammars- known as the Inside-Outside algorithm- has been found in practice to be "computationally intractable for realistic problems" [1]. Unsupervised learning algorithms have been designed for other grammar models (e.g. , [2, 3]). However, to the best of our knowledge, no large-scale experiments have been carried out to test the efficacy of these algorithms; the most likely reason is that their computational complexity, like that of the Inside-Outside algorithm, is impractical. One way to improve the complexity of inference and learning in statistical models is to introduce independence assumptions; however, doing so increases the model's bias. It is natural to wonder how a simpler grammar model (that can be trained efficiently from raw text) would compare with conventional models (which make fewer independence assumptions, but which must be trained from labelled data) . Such a model would be a useful tool in domains where partial accuracy is valuable and large amounts of unlabelled data are available (e.g., Information Retrieval, Information Extraction, etc.) . In this paper, we present a probabilistic model of syntax that is based upon grammatical bigrams, i.e., syntactic relationships between pairs of words. We show how this model results from introducing independence assumptions into more conven- the quick brown fox jumps over the lazy dog Figure 1: An example parse; arrows are drawn from head words to their dependents. The root word is jumps; brown is a predependent (adjunct) of fox; dog is a postdependent (complement) of over. tional models; as a result, grammatical bigram models can be trained efficiently from raw text using an O(n 3 ) EM algorithm. We present the results of experiments that quantify the representational adequacy of the grammatical bigram model, its ability to generalize from labelled data, and its ability to induce syntactic structure from large amounts of raw text. 2 The Grammatical Bigram Model We first provide a brief introduction to the Dependency Grammar formalism used by the grammatical bigram model; then, we present the probability model and relate it to conventional models; finally, we sketch the EM algorithm for training the model. Details regarding the parsing and learning algorithms can be found in a companion technical report [4]. Dependency Grammar Formalism. 1 The primary unit of syntactic structure in dependency grammars is the dependency relationship, or link- a binary relation between a pair of words in the sentence. In each link, one word is designated the head, and the other is its dependent. (Typically, different types of dependency are distinguished, e.g, subject, complement, adjunct, etc.; in our simple model, no such distinction is made.) Dependents that precede their head are called pre dependents, and dependents that follow their heads are called postdependents. A dependency parse consists of a set of links that, when viewed as a directed graph over word tokens , form an ordered tree. This implies three important properties: 1. Every word except one (the root) is dependent to exactly one head. 2. The links are acyclic; no word is, through a sequence of links, dependent to itself. 3. When drawn as a graph above the sentence, no two dependency relations cross-a property known as projectivity or planarity. The planarity constraint ensures that a head word and its (direct or indirect) dependents form a contiguous subsequence of the sentence; this sequence is the head word's constituent. See Figure 1 for an example dependency parse. In order to formalize our dependency grammar model, we will view sentences as sequences of word tokens drawn from some set of word types. Let V = {tl' t2, ... , t M } be our vocabulary of M word types. A sentence with n words is therefore represented as a sequence S = (Wl, W2 , ... , w n ), where each word token Wi is a variable that ranges over V. For 1 :S i , j :S n , we use the notation (i,j) E L to express that Wj is a dependent of Wi in the parse L. IThe Dependency Grammar formalism described here (which is the same used in [5 , 6]) is impoverished compared to the sophisticated models used in Linguistics; refer to [7] for a comprehensive treatment of English syntax in a dependency framework. Because it simplifies the structure of our model , we will make the following three assumptions about Sand L (without loss of generality): (1) the first word WI of S is a special symbol ROOT E V; (2) the root of L is WI; and, (3) WI has only one dependent. These assumptions are merely syntactic sugar: they allow us to treat all words in the true sentence (i.e., (W2, ... ,W n )) as dependent to one word. (The true root of the sentence is the sole child of WI.) Probability Model. A probabilistic dependency grammar is a probability distribution P(S, L) where S = (WI,W2, .. . ,wn ) is a sentence, L is a parse of S, and the words W2, ... ,Wn are random variables ranging over V. Of course, S and L exist in high dimensional spaces; therefore, tractable representations of this distribution make use of independence assumptions. Conventional probabilistic dependency grammar models make use of what may be called the head word hypothesis: that a head word is the sole (or primary) determinant of how its constituent combines with other constituents. The head word hypothesis constitutes an independence assumption; it implies that the distribution can be safely factored into a product over constituents: n P(S,L) = II P((Wj: (i,j) E L) is the dependent sequencelwi is the head) i=1 For example, the probability of a particular sequence can be governed by a fixed set of probabilistic phrase-structure rules , as in [6]; alternatively, the predependent and postdependent subsequences can be modeled separately by Markov chains that are specific to the head word, as in [8]. Consider a much stronger independence assumption: that all the dependents of a head word are independent of one another and their relative order. This is clearly an approximation; in general, there will be strong correlations between the dependents of a head word. More importantly, this assumption prevents the model from representing important argument structure constraints. (For example: many words require dependents , e.g. prepositions; some verbs can have optional objects, whereas others require or forbid them.) However, this assumption relieves the parser of having to maintain internal state for each constituent it constructs, and therefore reduces the computational complexity of parsing and learning. We can express this independence assumption in the following way: first , we forego modeling the length of the sentence, n, since in parsing applications it is always known; then, we expand P(S, Lin) into P(S I L)P(L I n) and choose P(L I n) as uniform; finally, we select II P(S I L) P( Wj is a [pre/post]dependent I Wi is the head) (i ,j)EL This distribution factors into a product of terms over syntactically related word pairs; therefore, we call this model the "grammatical bigram" model. The parameters of the model are <"(xy P(predependent is ty I head is t x ) 6. "(~ P(postdependent is ty Ihead is t x ) We can make the parameterization explicit by introducing the indicator variable wi, whose value is 1 if Wi = tx and a otherwise. Then we can express P(S I L) as P(S IL) (i,j)EL x=1 y=1 j<i (i,j)EL x=1 y=1 i<j Parsing. Parsing a sentence S consists of computing L* L:, argmaxP(L I S,n) = argmaxP(L, Sin) = argmaxP(S I L) L L L Yuret has shown that there are exponentially many parses of a sentence with n words [9], so exhaustive search for L * is intractable. Fortunately, our grammar model falls into the class of "Bilexical Grammars" , for which efficient parsing algorithms have been developed. Our parsing algorithm (described in the tech report [4]) is derived from Eisner's span-based chart-parsing algorithm [5], and can find L* in O(n 3 ) time. Learning. Suppose we have a labelled data set where Sk = (Wl,k, W2,k,?? ? , Wnk,k) and Lk is a parse over Sk. likelihood values for our parameters given the training data are The maximum et where the indicator variable is equal to 1 if (i,j) E Lk and 0 otherwise. As one would expect, the maximum-likelihood value of ,;; (resp. ,~ ) is simply the fraction of tx's predependents (resp. postdependents) that were ty. In the unsupervised acquisition problem, our data set has no parses; our approach is to treat the Lk as hidden variables and to employ the EM algorithm to learn (locally) optimal values of the parameters ,. As we have shown above, the are sufficient statistics for our model; the companion tech report [4] gives an adaptation of the Inside-Outside algorithm which computes their conditional expectation in O(n 3 ) time. This algorithm effectively examines every possible parse of every sentence in the training set and calculates the expected number of times each pair of words was related syntactically. et 3 Evaluation This section presents three experiments that attempt to quantify the representational adequacy and learnability of grammatical bigram models. Corpora. Our experiments make use of two corpora; one is labelled with parses, and the other is not. The labelled corpus was generated automatically from the phrase-structure trees in the Wall Street Journal portion of the Penn Treebank-III [10].2 The resultant corpus, which we call C, consists of 49,207 sentences (1,037,374 word tokens). This corpus is split into two pieces: 90% of the sentences comprise corpus Ctrain (44,286 sentences, 934,659 word tokens), and the remaining 10% comprise Ctest (4,921 sentences, 102,715 word tokens). The unlabelled corpus consists of the 1987- 1992 Wall Street Journal articles in the TREC Text Research Collection Volumes 1 and 2. These articles were segmented on sentence boundaries using the technique of [11], and the sentences were postprocessed to have a format similar to corpus C. The resultant corpus consists of 3,347,516 sentences (66,777,856 word tokens). We will call this corpus U. 2This involved selecting a head word for each constituent, for which the head-word extraction heuristics described in [6] were employed. Additionally, punctuation was removed, all words were down-cased, and all numbers were mapped to a special <#> symbol. The model's vocabulary is the same for all experiments; it consists of the 10,000 most frequent word types in corpus U; this vocabulary covers 94.0% of word instances in corpus U and 93.9% of word instances in corpus L. Words encountered during testing and training that are outside the vocabulary are mapped to the <unk> type. Performance metric. The performance metric we report is the link precision of the grammatical bigram model: the fraction of links hypothesized by the model that are present in the test corpus Ltest. (In a scenario where the model is not required to output a complete parse, e.g., a shallow parsing task, we could similarly define a notion of link recall; but in our current setting, these metrics are identical.) Link precision is measured without regard for link orientation; this amounts to ignoring the model's choice of root, since this choice induces a directionality on all of the edges. Experiments. We report on the results of three experiments: I. Retention. This experiment represents a best-case scenario: the model is trained on corpora Ltrain and Ltest and then tested on Ltest. The model's link precision in this setting is 80.6%. II. Generalization. In this experiment, we measure the model's ability to generalize from labelled data. The model is trained on Ltrain and then tested on Ltest. The model's link precision in this setting is 61.8%. III. Induction. In this experiment, we measure the model 's ability to induce grammatical structure from unlabelled data. The model is trained on U and then tested on Ltest . The model's link precision in this setting is 39.7%. Analysis. The results of Experiment I give some measure of the grammatical bigram model's representational adequacy. A model that memorizes every parse would perform perfectly in this setting, but the grammatical bigram model is only able to recover four out of every five links. To see why, we can examine an example parse. Figure 2 shows how the models trained in Experiments I, II, and III parse the same test sentence. In the top parse, syndrome is incorrectly selected as a postdependent of the first on token rather than the second. This error can be attributed directly to the grammatical bigram independence assumption: because argument structure is not modeled, there is no reason to prefer the correct parse, in which both on tokens have a single dependent , over the chosen parse, in which the first has two dependents and the second has none. 3 Experiment II measures the generalization ability of the grammatical bigram model; in this setting, the model can recover three out of every five links. To see why the performance drops so drastically, we again turn to an example parse: the middle parse in Figure 2. Because the forces -+ on link was never observed in the training data, served has been made the head of both on tokens; ironically, this corrects the error made in the top parse because the planarity constraint rules out the incorrect link from the first on token to syndrome. Another error in the middle parse is a failure to select several as a predependent of forces; this error also arises because the combination never occurs in the training data. Thus, we can attribute this drop in performance to sparseness in the training data. We can compare the grammatical bigram model's parsing performance with the results reported by Eisner [8]. In that investigation, several different probability models are ascribed to the simple dependency grammar described above and 3 Although the model's parse of acquired immune deficiency syndrome agrees with the labelled corpus, this particular parse reflects a failure of the head-word extraction heuristics; acquired and immune should be predependents of deficiency, and deficiency should be a predependent of syndrome . 1. 843 '. 88' I. fi r hr-. n~ 14 . 383 <root> she has also served on several task forces on acquired immune deficiency syndrome r,m fO ~8~\ fi r hn ' . 803 II. 3.2",-d r 1r 1. 528 1.358 A 9 . 630 , 1 1 2 . 527 14 . 264 - <root> she has also served on several task forces on acquired immune deficiency syndrome 1. 990 0.913 III. tI k 4 . 124 <root> she has also served 0 .14 9 - 1.709 1--' on several f' (~ 13 . 585 ~( task forces on acquired immune deficiency syndrome Figure 2: The same test sentence, parsed by the models trained in each of the three experiments. Links are labelled with -log2 IXY I I:~1 IXY, the mutual information of the linked words; dotted edges are default attachments. are compared on a task similar to Experiment 11.4 Eisner reports that the bestperforming dependency grammar model (Model D) achieves a (direction-sensitive) link precision of 90.0%, and the Collins parser [6] achieves a (direction-sensitive) link precision of 92.6%. The superior performance of these models can be attributed to two factors: first, they include sophisticated models of argument structure; and second, they both make use of part-of-speech taggers, and can "back-off" to non-lexical distributions when statistics are not available. Finally, Experiment III shows that when trained on unlabelled data, the grammatical bigram model is able to recover two out of every five links. This performance is rather poor, and is only slightly better than chance; a model that chooses parses uniformly at random achieves 31.3% precision on L\est . To get an intuition for why this performance is so poor, we can examine the last parse, which was induced from unlabelled data. Because Wall Street Journal articles often report corporate news, the frequent co-occurrence of has -+ acquired has led to a parse consistent with the interpretation that the subject she suffers from AIDS, rather than serving on a task force to study it. We also see that a flat parse structure has been selected for acquired immune deficiency syndrome; this is because while this particular noun phrase occurs in the training data, its constituent nouns do not occur independently with any frequency, and so their relative co-occurrence frequencies cannot be assessed. 4 Discussion Future work. As one would expect, our experiments indicate that the parsing performance of the grammatical bigram model is not as good as that of state-ofthe-art parsers; however, its performance in Experiment II suggests that it may be useful in domains where partial accuracy is valuable and large amounts of unlabelled data are available. However, to realize that potential, the model must be improved so that its performance in Experiment III is closer to that of Experiment II. To that end, we can see two obvious avenues of improvement. The first involves increasing the model's capacity for generalization and preventing overfitting. The 4The labelled corpus used in that investigation is also based upon a transformed version of Treebank-III, but the head-word extraction heuristics were slightly different , and sentences with conjunctions were completely eliminated. However, the setup is sufficiently similar that we think the comparison we draw is informative. model presented in this paper is sensitive only to pairwise relationships to words; however, it could make good use of the fact that words can have similar syntactic behavior. We are currently investigating whether word clustering techniques can improve performance in supervised and unsupervised learning. Another way to improve the model is to directly address the primary source of parsing error: the lack of argument structure modeling. We are also investigating approximation algorithms that reintroduce argument structure constraints without making the computational complexity unmanageable. Related work. A recent proposal by Yuret presents a "lexical attraction" model with similarities to the grammatical bigram model [9]; however, unlike the present proposal, that model is trained using a heuristic algorithm. The grammatical bigram model also bears resemblance to several proposals to extend finite-state methods to model long-distance dependencies (e.g., [12, 13]), although these models are not based upon an underlying theory of syntax. References [1] K. Lari and S. J. Young. The estimation of stochastic context-free grammars using the Inside-Outside algorithm. Computer Speech and Language, 4:35- 56, 1990. [2] John Lafferty, Daniel Sleator, and Davy Temperley. Grammatical trigrams: A probabilistic model of link grammar. In Proceedings of the AAAI Conference on Probabilistic Approaches to Natural Language, October 1992. [3] Yves Schabes. Stochastic lexicalized tree-adjoining grammars. In Proceedings of the Fourteenth International Conference on Computational Linguistics, pages 426-432, Nantes, France, 1992. [4] Mark A. Paskin. Cubic-time parsing and learning algorithms for grammatical bigram models. Technical Report CSD-01-1148, University of California, Berkeley, 2001. [5] Jason Eisner. Bilexical grammars and their cubic-time parsing algorithms. In Harry Bunt and Anton Nijholt, editors, Advances in Probabilistic and Other Parsing Technologies, chapter 1. Kluwer Academic Publishers, October 2000. [6] Michael Collins. Head-driven Statistical Models for Natural Language Parsing. PhD thesis, University of Pennsylvania, Philadelphia, Pennsylvania, 1999. [7] Richard A. Hudson. English Word Grammar. B. Blackwell, Oxford, UK, 1990. [8] Jason M. Eisner. An empirical comparison of probability models for dependency grammars. Technical Report ICRS-96-11, CIS Department, University of Pennsylvania, 220 S. 33 rd St. Philadelphia, PA 19104- 6389, 1996. [9] Deniz Yuret . Discovery of Linguistic R elations Using Lexical Attraction. PhD thesis, Massachusetts Institute of Technology, May 1998. [10] M. Marcus, B. Santorini, and M. Marcinkiewicz. Building a large annotated corpus of english: The penn treebank. Computational Linguistics, 19:313- 330, 1993. [11] Jeffrey C. Reynar and Adwait Ratnaparkhi. A maximum entropy approach to identifying sentence boundaries. In Proceedings of the Fifth Conference on Appli ed Natural Language Processing, Washington, D.C. , March 31 - April 3 1997. [12] S. Della Pietra, V. Della Pietra, J. Gillett, J. Lafferty, H. Printz , and L. Ures. Inference and estimation of a long-range trigram model. In Proceedings of the Second International Colloquium on Grammatical Inference and Applications, number 862 in Lecture Notes in Artificial Intelligence, pages 78- 92. Springer-Verlag, 1994. [13] Ronald Rosenfeld. Adaptive Statistical Language Modeling: A Maximum Entropy Approach. PhD thesis, Carnegie Mellon University, 1994.
2041 |@word determinant:1 middle:2 manageable:1 bigram:21 stronger:1 version:1 efficacy:1 selecting:1 daniel:1 current:1 must:2 parsing:16 john:1 realize:1 ronald:1 realistic:1 deniz:1 informative:1 designed:1 drop:2 intelligence:1 fewer:1 selected:2 parameterization:1 simpler:2 five:3 ironically:1 tagger:1 direct:1 incorrect:1 consists:6 combine:1 inside:4 introduce:1 acquired:7 pairwise:1 ascribed:1 expected:1 behavior:1 examine:2 automatically:1 increasing:1 notation:1 underlying:1 what:1 developed:1 impractical:1 safely:1 berkeley:4 every:7 ti:1 exactly:1 uk:1 unit:1 penn:2 retention:1 hudson:1 treat:2 bilexical:2 adwait:1 oxford:1 planarity:3 suggests:1 co:2 range:2 directed:1 testing:1 practice:1 empirical:1 davy:1 word:48 induce:3 pre:2 get:1 cannot:1 ctest:1 context:2 applying:1 conventional:4 quick:1 lexical:3 independently:1 identifying:1 factored:1 rule:2 yuret:3 examines:1 importantly:1 attraction:2 notion:1 resp:2 suppose:1 parser:3 hypothesis:2 pa:1 observed:1 wj:3 ensures:1 news:1 reintroduce:1 removed:1 valuable:2 projectivity:1 intuition:1 colloquium:1 complexity:6 sugar:1 trained:10 ithe:1 upon:4 division:1 completely:1 indirect:1 represented:1 tx:2 chapter:1 artificial:1 outside:5 exhaustive:1 whose:1 heuristic:4 ratnaparkhi:1 otherwise:2 grammar:23 ability:7 statistic:2 schabes:1 syntactic:7 think:1 itself:1 rosenfeld:1 sequence:5 product:2 adaptation:1 frequent:2 representational:4 constituent:7 argmaxp:3 postprocessed:1 object:1 measured:1 sole:2 strong:1 c:1 involves:1 implies:2 indicate:1 quantify:3 direction:2 correct:1 attribute:1 annotated:1 stochastic:2 appli:1 sand:1 require:2 generalization:3 wall:3 investigation:2 marcinkiewicz:1 sufficiently:1 trigram:2 achieves:3 estimation:2 precede:1 currently:1 sensitive:3 agrees:1 wl:2 tool:1 reflects:1 clearly:1 always:1 rather:3 conjunction:1 linguistic:1 derived:2 she:4 improvement:1 likelihood:2 tech:2 inference:3 tional:1 dependent:18 el:3 typically:1 hidden:1 relation:2 expand:1 transformed:1 nantes:1 france:1 unk:1 orientation:1 noun:2 special:2 art:1 mutual:1 equal:1 construct:1 comprise:2 extraction:4 having:1 never:2 eliminated:1 identical:1 represents:1 washington:1 unsupervised:5 constitutes:1 future:1 report:9 t2:1 others:1 richard:1 employ:1 comprehensive:1 pietra:2 relief:1 jeffrey:1 maintain:1 attempt:1 evaluation:1 punctuation:1 adjoining:1 chain:1 edge:2 closer:1 partial:2 xy:1 fox:2 tree:3 bestperforming:1 instance:2 formalism:3 modeling:3 contiguous:1 cover:1 phrase:3 introducing:2 uniform:1 wonder:1 learnability:1 reported:1 dependency:17 chooses:1 st:1 international:2 forbid:1 probabilistic:9 off:1 corrects:1 michael:1 again:1 aaai:1 thesis:3 choose:1 hn:1 forego:1 potential:1 harry:1 piece:1 root:9 jason:2 view:1 memorizes:1 doing:1 linked:1 portion:1 recover:3 il:1 chart:1 accuracy:2 yves:1 efficiently:2 ofthe:1 generalize:3 anton:1 raw:5 none:1 served:4 fo:1 suffers:1 ed:1 failure:2 ty:3 acquisition:1 frequency:2 involved:1 cased:1 obvious:1 resultant:2 attributed:2 treatment:1 massachusetts:1 recall:1 knowledge:1 formalize:1 impoverished:1 sophisticated:2 back:1 supervised:1 follow:1 improved:1 april:1 generality:1 correlation:1 sketch:1 parse:23 lack:1 resemblance:1 building:1 hypothesized:1 brown:2 true:2 sin:1 during:1 ixy:2 syntax:3 complete:1 syntactically:2 ranging:1 fi:2 superior:1 ltrain:2 exponentially:1 volume:1 extend:1 interpretation:1 kluwer:1 significant:1 refer:1 mellon:1 rd:1 similarly:1 adjunct:2 language:5 immune:6 similarity:1 etc:2 recent:1 driven:1 scenario:2 verlag:1 binary:1 fortunately:1 employed:1 syndrome:8 ii:8 corporate:1 reduces:1 segmented:1 technical:3 unlabelled:6 academic:1 cross:1 long:2 retrieval:1 lin:1 post:1 calculates:1 expectation:1 metric:3 proposal:3 whereas:1 separately:1 ures:1 source:1 publisher:1 printz:1 w2:5 unlike:1 subject:2 induced:1 lafferty:2 adequacy:4 call:3 iii:7 split:1 wn:2 independence:8 pennsylvania:3 perfectly:1 regarding:1 simplifies:1 avenue:1 whether:1 speech:2 useful:2 amount:5 locally:1 induces:1 exist:1 dotted:1 serving:1 carnegie:1 express:3 four:1 drawn:3 graph:2 merely:1 fraction:2 fourteenth:1 draw:1 prefer:1 encountered:1 occur:1 constraint:4 deficiency:7 flat:1 argument:5 span:1 format:1 department:1 designated:1 combination:1 poor:2 march:1 slightly:2 em:5 wi:10 shallow:1 making:1 sleator:1 computationally:1 lari:1 turn:1 tractable:1 end:1 available:3 occurrence:2 distinguished:1 top:2 remaining:1 linguistics:3 include:1 clustering:1 log2:1 parsed:1 eisner:5 wnk:1 occurs:2 primary:3 distance:1 link:22 mapped:2 capacity:1 street:3 reason:2 induction:1 marcus:1 length:1 modeled:2 relationship:4 setup:1 october:2 relate:1 perform:1 markov:1 finite:1 optional:1 incorrectly:1 santorini:1 head:22 trec:1 verb:1 complement:2 pair:5 dog:2 required:1 blackwell:1 sentence:23 california:2 distinction:1 bunt:1 address:1 able:2 challenge:1 natural:4 force:6 indicator:2 hr:1 representing:1 improve:3 technology:2 brief:1 lk:3 attachment:1 carried:1 philadelphia:2 text:6 discovery:1 ltest:5 relative:2 loss:1 par:4 expect:2 bear:1 lecture:1 acyclic:1 sufficient:1 consistent:1 article:3 treebank:3 editor:1 preposition:1 course:1 token:11 last:1 keeping:1 english:5 free:2 drastically:1 bias:1 allow:1 institute:1 fall:1 unmanageable:1 fifth:1 grammatical:24 boundary:2 regard:1 vocabulary:4 default:1 computes:1 preventing:1 made:3 jump:2 collection:1 adaptive:1 overfitting:1 investigating:2 corpus:18 alternatively:1 subsequence:2 search:1 sk:2 why:3 additionally:1 learn:1 ca:1 ignoring:1 domain:2 csd:1 arrow:1 elation:1 child:1 tl:1 cubic:2 aid:1 precision:8 explicit:1 governed:1 young:1 companion:2 down:1 specific:1 paskin:3 symbol:2 admits:1 intractable:3 lexicalized:1 effectively:1 ci:1 phd:3 sparseness:1 entropy:2 led:1 simply:1 likely:1 lazy:1 prevents:1 ordered:1 springer:1 chance:1 conditional:1 viewed:1 labelled:10 directionality:1 except:1 uniformly:1 called:3 est:1 select:2 internal:1 mark:2 arises:1 collins:2 assessed:1 tested:3 della:2
1,143
2,042
Boosting and Maximum Likelihood for Exponential Models Guy Lebanon School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 John Lafferty School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 lebanon@cs.cmu.edu lafferty@cs.cmu.edu Abstract We derive an equivalence between AdaBoost and the dual of a convex optimization problem, showing that the only difference between minimizing the exponential loss used by AdaBoost and maximum likelihood for exponential models is that the latter requires the model to be normalized to form a conditional probability distribution over labels. In addition to establishing a simple and easily understood connection between the two methods, this framework enables us to derive new regularization procedures for boosting that directly correspond to penalized maximum likelihood. Experiments on UCI datasets support our theoretical analysis and give additional insight into the relationship between boosting and logistic regression. 1 Introduction Several recent papers in statistics and machine learning have been devoted to the relationship between boosting and more standard statistical procedures such as logistic regression. In spite of this activity, an easy-to-understand and clean connection between these different techniques has not emerged. Friedman, Hastie and Tibshirani [7] note the similarity between boosting and stepwise logistic regression procedures, and suggest a least-squares alternative, but view the loss functions of the two problems as different, leaving the precise relationship between boosting and maximum likelihood unresolved. Kivinen and Warmuth [8] note that boosting is a form of ?entropy projection,? and Lafferty [9] suggests the use of Bregman distances to approximate the exponential loss. Mason et al. [10] consider boosting algorithms as functional gradient descent and Duffy and Helmbold [5] study various loss functions with respect to the PAC boosting property. More recently, Collins, Schapire and Singer [2] show how different Bregman distances precisely account for boosting and logistic regression, and use this framework to give the first convergence proof of AdaBoost. However, in this work the two methods are viewed as minimizing different loss functions. Moreover, the optimization problems are formulated in terms of a reference distribution consisting of the zero vector, rather than the empirical distribution of the data, making the interpretation of this use of Bregman distances problematic from a statistical point of view. In this paper we present a very basic connection between boosting and maximum likelihood for exponential models through a simple convex optimization problem. In this setting, it is seen that the only difference between AdaBoost and maximum likelihood for exponential models, in particular logistic regression, is that the latter requires the model to be normalized to form a probability distribution. The two methods minimize the same extended Kullback-Leibler divergence objective function subject to the same feature constraints. Using information geometry, we show that projecting the exponential loss model onto the simplex of conditional probability distributions gives precisely the maximum likelihood exponential model with the specified sufficient statistics. In many cases of practical interest, the resulting models will be identical; in particular, as the number of features increases to fit the training data the two methods will give the same classifiers. We note that throughout the paper we view boosting as a procedure for minimizing the exponential loss, using either parallel or sequential update algorithms as in [2], rather than as a forward stepwise procedure as presented in [7] or [6]. Given the recent interest in these techniques, it is striking that this connection has gone unobserved until now. However in general, there may be many ways of writing the constraints for a convex optimization problem, and many different settings of the Lagrange multipliers (or Kuhn-Tucker vectors) that represent identical solutions. The key to the connection we present here lies in the use of a particular non-standard presentation of the constraints. When viewed in this way, there is no need for special-purpose Bregman distances to give a unified account of boosting and maximum likelihood, as we only make use of the standard Kullback-Leibler divergence. But our analysis gives more than a formal framework for understanding old algorithms; it also leads to new algorithms for regularizing AdaBoost, which is required when the training data is noisy. In particular, we derive a regularization procedure for AdaBoost that directly corresponds to penalized maximum likelihood using a Gaussian prior. Experiments on UCI data support our theoretical analysis, demonstrate the effectiveness of the new regularization method, and give further insight into the relationship between boosting and maximum likelihood exponential models. 2 Notation              ( '*),+.-0!/ #"%$&(('*'*-&),+.45-0)6/&/ 1%+  2)   3 798 :   <; =1>+@??@?6+. LMN'B)O+.-0/ MPL MNL'*)D'*/ )6/XWZY MNL 'B)O+.-0/Q)  HK  A'BHCRI,)DC.KT+ES -%'*CF)U/GCV+EHCJI,)D/K S 'B-%C.+E-0/G? )[ -TL '*)6/ MNL '*)6/QWY -\] MNL '*-&45)6/PW^Y ) ) Let and be finite sets. We denote by the set of nonnegative measures on , and by the set of conditional probability distributions, for each . For , we will overload the notation and ; the latter will be suggestive of a conditional probability distribution, but in general it need not be normalized. Let , , be given functions, which we will refer to as features. These will correspond to the weak learners in boosting, and to the sufficient statistics in an exponential model. Suppose that we have data with empirical distribution and marginal ; thus, We assume, without loss of generality, that for all . Throughout the paper, we assume (for notational convenience) that the training data has the following property. Consistent Data Assumption. For each with for which . This will be denoted . , there is a unique For most data sets of interest, each appears only once, so that the assumption trivially holds. However, if appears more than once, we require that it is labeled consistently. We make this assumption mainly to correspond with the conventions used to present boosting algorithms; it is not essential to what follows. g<hUikjRlK m nV7 oq8 prm h.sRt9u>v `%w xzyJ{zw r|F} Given ~FaO+V7c'*),+.-0/.^ #_ `D'*-&8b I,45K)Da/ 8 7 8 '*),a3+.-0/ 3cb a , we define the exponential model where hood estimation problem is to determine parameters , for _`d'B-4e)D/f , by . The maximum likeli- that maximize the conditional log- %'FaD/  {zw LMN'B)O+.-0/ N_`U'*-&45)6/ %' aU/ 'FaD/]  HCJI,K Z I ucv `%w x%yR{ w r| %x yR{ w *|B} ? likelihood or minimize the log loss  . The objective  function to be minimized in the multi-label boosting algorithm AdaBoost.M2 [2] is the ex  ponential loss given by M2 As has been often noted, the log loss and the exponential loss are qualitatively different. The exponential loss grows exponentially with increasing negative ?margin,? while the log loss grows linearly. 3 Correspondence Between AdaBoost and Maximum Likelihood Since we are working with unnormalized models we make use of the extended conditional Kullback-Leibler divergence or  -divergence, given by ',+E_z/   { MQL '*)D/  ,'B- 45)6/  _ '*'*--4e45)D)D//  '*-&45)D/ _ 'B-4e)D/    ,'#45)6/  _ '#45)6/  78  _   ]'*MPL +V7T/N   '*MNL +k7T/  ! [   { MNL 'B)D/  '*-45)D/ 'F7 8 'B)O+.-0/"$#%&(' 7 8 45)*) /  Y0+ ;d>? ML + #,%&(' 7 45-) )X 7c'B)O+V-6L 'B)D/E/ . K .0/   '1. / / ']43O+V_ '* ML / +V7T/ 1' . K@/ ' O+V_  / [2]' MN L +k7T/    . / . K .6K 5   defined on features def (possibly taking on the value ). Note that if and then this becomes the more familiar KL divergence for probabilities. Let and a fixed default distribution be given. We define as all (1) , this set is non-empty. Note that under the consistent data assumption, we Since have that . Consider now the following two convex optimization problems, labeled and . minimize subject to minimize subject to Thus, problem differs from only in that the solution is required to be normalized. As we?ll show, the dual problem corresponds to AdaBoost, and the dual problem .7/ 5 corresponds to maximum likelihood for exponential models.  { MQL '*)D/ Z ,'B- 4e)D/07c'B)O+.-0/& This presentation of the constraints is the key to making the correspondence between AdaBoost and maximum likelihood. Note that the constraint  #%&0' 8) , which is the usual presentation of the constraints for maximum likelihood (as dual to maximum entropy), doesn?t make sense for unnormalized models, since the two sides of the equation may not be ?on the same scale.? Note further that attempting to rescale by dividing by the mass of  to get 7  M{ PL 'B)D/  Z,'B-9,4e)D'*-&/07c45)6'B)O/ +E- / :#%&0' 78) would yield nonlinear constraints. K )D/ , B '  e 4 K '3O+kaU/  { LMN'*)6/ ,'B- 4e)D/ <  _  '*-&45)D/  1,^~FaO+V7c'B)O+.-0/"$#%&0' 7<45)*)   ? a uA v b`%w x%yR{zw @|9D8EFHG x"I*{KJ } _#`  =?>@BA " C ; K9'3O; +VaD/ _`D'*-45)6/X _  '*-&45)6/ ?  L M6K9' aU/N % K B '  _ U ` V + D a / L M6K9'FaD/   M{ PL '*)6/ _  '*-45)D/ u v `%w xzyJ{zw r|9D8EFHG x"IB{KJ } ? a N P OQRW,S{ =TOVU ` LMO' aU/  8,'B-4e)D/ =1 . / 8 We now derive the dual problems formally; the following section gives a precise statement of the duality result. To derive the dual problem .75 , we calculate the Lagrangian as ; For     def , the connecting equation arg Thus, the dual function is given by is given by (2) The dual problem is to determine simply add additional Lagrange multipliers . To derive the dual for for the constraints . , we 3.1 Special cases It is now straightforward to derive various boosting and logistic regression problems as special cases of the above optimization problems. _ A'B- 4e)D/   1 uAv `zw x%yR{ w r| x%yR{ w *| =TO U `LM K 'FaD/ ` C I - 2 &1%+ u  61%  v `%w x%yR{ B|B} 7 8 'B)O+E- / /K -N7 8 'B)D/ a N  OQR =7>@ `  C 1%+   { 1>MNL+@?@'*?)6?T/ +  ?  '*-&45)6/.7 8 'B)O+.-0/   {zw MNL 'B)O+.-0/E7 8 '*),+.-0/  8,'B-4e) C / _u>` v `%'*-&w xzyJ45{z)6w r|/ F}  K p K_ >'B-4e)D/ uAv `zw x%yR{%w r|B} { W { {  Zc_ >'*-4 )D/ L '*)O+E-0/ <  _`D'B-4e)D/ L M / ' aU/<  { MQ 798>'*)O+E-0/ a  Case 1: AdaBoost.M2. Take  lent to computing *N OVQ! =?><@ problem of AdaBoost.M2. . Then the dual problem is equivawhich is the optimization Case 2: Binary AdaBoost. In addition to the assumptions for the previous case, now assume that , and take . Then the dual problem is given by which is the optimization problem of binary AdaBoost. Case 3: Maximum Likelihood for Exponential Models. In this case we take the same setup as for AdaBoost.M2 but add the additional normalization constraints:   If these constraints are satisfied, then the other constraints take the form and the connecting  equation becomes were  is the normalizing term  , which corresponds to setting the Lagrange multiplier to the appropriate value. In this case, after a simple calculation the dual problem is seen to be which corresponds to maximum likelihood for a conditional exponential model with sufficient statistics . _`D'e1 4e)D/& K  i jRK l m nEo p sJt ? Case 4: Logistic Regression. Returning to the case of binary AdaBoost, we see that when we add normalization constraints as above, the model is equivalent to binary logistic regres sion, since We note that it is not necessary to scale the features by a constant factor here, as in [7]; the correspondence between logistic regression and boosting is direct. 3.2 Duality K K 'F_K%+k7T/  # _ [ &_ '*-&45)6/  _ A'*-&45)6/ u u v `%v `%w xzw x%yJ{zyR{zw r|w @|F} x%yR{%w yR{ |B|B} +Ua][ b  /%'F_K%+k7T/  # _   &_ '*-&45)6/ _ A'*-&45)D/ +Ua][ b A? / _ N K _N _ N  OQR =7>@ #" T { MNL '*)6/d c_ '*-&45)D/ _ N :OVQ!S=TO U #" O { MPL '*)D/B<  _ ' -&45)6/    ' Y0+V_z/  '*MNL +E_z/  ' MNL +V_  /  _ N _ N _ N  OVQ!A  " =?><@  3' O+V_  /  OVQ!S" =7>@  '*MNL +E_z/ _ N  OVA Q!"  =?><@  3' O+VK_ /  OVQ!S" =7>@  '*MNL + /6? _N _ N _ N : OVQ! =?>@ & "  ' O+V_ N / Let    and  / be defined as the following exponential families:  K K is unnormalized while  is normalized. We now define the boosting solution and maximum likelihood solution ml as boost Thus  boost    ml   where  denotes the closure of the set  . The following theorem corresponds to Proposition 4 of [3] for the usual KL divergence; in [4] the duality theorem is proved for a general class of Bregman distances, including the extended KL divergence as a special as in [2], but rather case. Note that we do not work with divergences such as  , which is more natural and interpretable from a statistical point-of-view. Theorem. Suppose that boost ml  . Then boost and ml exist, are unique, and satisfy     !    Moreover, ml is computed in terms of boost as ml #" ! boost . K K  _ PSfrag replacements  PSfrag replacements N boost   _N _N boost ml ml  Figure 1: Geometric view of the duality theorem. Minimizing the exponential loss finds the member    that intersects with the feasible set of measures satisfying the moment constraints (left). When of we impose the additional constraint that each conditional distribution must be normalized, we introduce a Lagrange multiplier for each training example , giving a higher-dimensional family . By the duality theorem, projecting the exponential loss solution onto the intersection of the feasible set with the simplex gives the maximum likelihood solution.   K This result has a simple geometric interpretation. The unnormalized exponential family intersects the feasible set of measures satisfying the constraints (1) at a single point.  The algorithms presented  in [2] determine this point, which is the exponential loss solution Nboost :OVQ!S=?><@  (see Figure 1, left). _  '*ML +E_z/ " _  "  ' ML +V_z/ K On the other hand, maximum conditional likelihood estimation for an exponential model  with the same features is equivalent to the problem Vml where  N OQRS=?>@    is the exponential family with additional Lagrange multipliers, one for each normalization constraint. The feasible set for this problem is . Since , by the Pythagorean (see Figure 1, right). equality we have that ml N OQR =7>@ A ! 3 boost N _  " ' O+V _ /    4 Regularization 78 Minimizing the exponential loss or the log loss on real data often fails to produce finite parameters. Specifically, this happens when for some feature or 7 8 '*),+.-0/"7 8 '*),+G-L '*)6/./^Y 7 8 '*),+.-0"/ 7 8 '*),+G-L '*)6/./ ^Y for all for all - ) - ) and and with with MPL 'B)D/PWY MPL 'B)D/PWY0? (3) This is especially harmful since often the features for which (3) holds are the most important for the purpose of discrimination. Of course, even when (3) does not hold, models trained by maximum likelihood or the exponential loss can overfit the training data. A standard regularization technique in the case of maximum likelihood employs parameter priors in a Bayesian framework. See [11] for non-Bayesian alternatives in the context of boosting. a In terms of convex duality, parameter priors for the dual problem correspond to ?potentials? on the constraint values in the primal problem. The case of a Gaussian prior on , for example, corresponds to a quadratic potential on the constraint values in the primal problem. ' O+ / *' MNL ^+k^7D+ @/ N^ \ b ]' ML +k7D+ @/  [    { MQL '*)D/   '*-&45)6/'F78>'B)O+.-0/"$#%&0' 78 45)*) /  V8 . Kkw  ' .cKGw / ' ,+E_  /  ' @/ [4 '*ML +V7D+ @/ f% b  Y ; ; ' ,+ 9+kaU/& ' ,+VaD/S ' @/ M Kkw 'FaD/  M K 'FaD( /  K / 5z/'FaD/ 5%'FaD/  / ' @/   8 / 8 8 5%'FaD/X   8 /K 8 a 8/  M6KGw ' aU/   L{ MN'B)D/  _  '*-&45)6/ u v `%w x%yR{%w r| x%yR{%w yR{ |B}   8 a 8/ 8/ ? We now consider primal problems over 3  where  vector that relaxes the original constraints. Define  and  is a parameter as    and consider the primal problem reg where  (4) reg given by  minimize subject to  is a convex function whose minimum is at .  To derive the dual problem, the Lagrangian is calculated as  and the dual function is reg where  , we have  of  . For a quadratic penalty   the dual function becomes  reg  is the convex conjugate and   (5) A sequential update rule for (5) incurs the small additional cost of solving a nonlinear equation by Newton-Raphson every iteration. See [1] for a discussion of this technique in the context of exponential models in statistical language modeling. 5 Experiments We performed experiments on some of the UC Irvine datasets in order to investigate the relationship between boosting and maximum likelihood empirically. The weak learner was the decision stump FindAttrTest as described in [6], and the training set consisted of a randomly chosen 90% of the data. Table 1 shows experiments with regularized boosting. Two boosting models are compared. The first model was trained for 10 features generated by FindAttrTest, excluding features satisfying condition (3). Training was carried out using the parallel update method described in [2]. The second model, / , was trained using the exponential loss with quadratic regularization. The performance was measured using the conditional log-likelihood of the (normalized) models over the training and test set, denoted train and test , as well as using the test error rate  test . The table entries were averaged by 10-fold cross validation. _zK _ For the weak learner FindAttrTest, only the Iris dataset produced features that satisfy (3). On average, 4 out of the 10 features were removed. As the flexibility of the weak learner is increased, (3) is expected to hold more often. On this dataset regularization improves both the test set log-likelihood and error rate considerably. In datasets where shows significant overfitting, regularization improves both the log-likelihood measure and the error rate. In cases of little overfitting (according to the log-likelihood measure), regularization only improves the test set log-likelihood at the expense of the training set log-likelihood, however without affecting test set error. _K _ _ differs from ml Next we performed a set of experiments to test how much boost N N , where the boosting model is normalized (after training) to form a conditional probability distribution. For different experiments, FindAttrTest generated a different number of features (10?100), and the training set was selected randomly. The top row in Figure 2 shows for the Sonar dataset as well as between N and train boost N  the relationship between train ml N and train ml N Nboost . As the number of features increases so that the training train ml 'F_ / 'B_ +V_ / 'F_ / 'F_ / Data Promoters Iris Sonar Glass Ionosphere Hepatitis Breast Pima train 'F_K/ 'B_#K@/ Unregularized test -0.29 -0.29 -0.22 -0.82 -0.18 -0.28 -0.12 -0.48 -0.60 -1.16 -0.58 -0.90 -0.36 -0.42 -0.14 -0.53  test 'F_K/ 0.28 0.21 0.25 0.36 0.13 0.19 0.04 0.26 train 'B_ / / -0.32 -0.10 -0.26 -0.84 -0.21 -0.28 -0.12 -0.48 'F_ / / Regularized test -0.50 -0.20 -0.48 -0.90 -0.28 -0.39 -0.14 -0.52  test 'B_ / / 0.26 0.09 0.19 0.36 0.10 0.19 0.04 0.25 Table 1: Comparison of unregularized to regularized boosting. For both the regularized and unregularized cases, the first column shows training log-likelihood, the second column shows test loglikelihood, and the third column shows test error rate. Regularization reduces error rate in some cases while it consistently improves the test set log-likelihood measure on all datasets. All entries were averaged using 10-fold cross validation. 'B_ / U Y data is more closely fit ( train ml 7 ), the boosting and maximum likelihood models become more similar, as measured by the KL divergence. This result does not hold when the model is unidentifiable and the two models diverge in arbitrary directions. 'B_ / 'F_ / 'B_ / 'F_ / The bottom row in Figure 2 shows the relationship between the test set log-likelihoods, , together with the test set error rates  test ml . In N and test boost N N and  test boost N test ml these figures the testing set was chosen to be 50% of the total data. In order to indicate the number of points at each error rate, each circle was shifted by a small random value to avoid points falling on top of each other. While the plots in the bottom row of Figure 2 indicate that train ml , as expected, on the test data the linear trend is reversed, N N train boost so that test ml N  test boost N . Identical experiments on Hepatitis, Glass and Promoters resulted in similar results and are omitted due to lack of space. 'F_ 'B_ / / W 'B'F_ _ / / The duality result suggests a possible explanation for the higher performance of boosting with respect to test . The boosting model is less constrained due to the lack of normalization constraints, and therefore has a smaller  -divergence to the uniform model. This may be interpreted as a higher extended entropy, or less concentrated conditional model. %'F_ / D Y _#`\ + 'B%'_ F_ +V_/ ` /  %'%'BB__ / / %'B_`>/r? ML However, as ml , the two models come to agree (up to identifiability). It is easy to N  show that for any exponential model By taking  / N N " train ml ml N it is seen that as the difference between Vml N and boost N gets smaller, the boost divergence between the two models also gets smaller. The empirical results are consistent with the theoretical analysis. As the number of features is increased so that the training data is fit more closely, the model matches the empirical distribution and the normalizing N becomes a constant. In this case, normalizing the boosting model " boost does term  not violate the constraints, and results in the maximum likelihood model. _`( _ ` 'B)D/ Acknowledgments We thank Michael Collins, Michael Jordan, Andrew Ng, Fernando Pereira, Rob Schapire, and Yair Weiss for helpful comments on an early version of this paper. Part of this work was carried out while the second author was visiting the Department of Statistics, University of California at Berkeley.  ?0.05  boost  ?0.1 train  boost   0.04  0.035 0.025 ml 0.045  train 0 0.03 ?0.15 0.02 ?0.2 0.015 0.01 PSfrag replacements ?0.25 PSfrag replacements ?0.3 ?0.3 ?0.25 ?0.2 ?0.15 train   ?0.1 ?0.05 0.005 0 ?0.25 0 ?0.2 ?0.15 ml  ?0.1 train  m  l ?0.05 0 0.4 ?10  0.35 0.3 test  test  boost ?5 boost    0.25 ?15 0.2 ?20 PSfrag replacements PSfrag replacements ?25 ?25 ?20   ?15 0.15 0.1 ?10 test m  l ?5     0.1     0.15 0.2       0.25   0.3 test 0.35 0.4 ml  Figure 2: Comparison of AdaBoost and maximum likelihood for Sonar dataset. The top row com-        boost  pares  train ml to  train boost (left) and  train ml to  train ml (right). The bottom row shows the relationship between  test ml and  test boost (left) and test ml and test boost (right).     The experimental results for other UCI datasets were very similar. References [1] S. Chen and R. Rosenfeld. A survey of smoothing techniques for ME models. IEEE Transactions on Speech and Audio Processing, 8(1), 2000. [2] M. Collins, R. E. Schapire, and Y. Singer. Logistic regression, AdaBoost and Bregman distances. Machine Learning, to appear. [3] S. Della Pietra, V. Della Pietra, and J. Lafferty. Inducing features of random fields. IEEE Transactions on Pattern Analysis and Machine Intelligence, 19(4), 1997. [4] S. Della Pietra, V. Della Pietra, and J. Lafferty. Duality and auxiliary functions for Bregman distances. Technical Report CMU-CS-01-109, Carnegie Mellon University, 2001. [5] N. Duffy and D. Helmbold. Potential boosters? In Neural Information Processing Systems, 2000. [6] Y. Freund and R. E. Schapire. Experiments with a new boosting algorithm. In International Conference on Machine Learning, 1996. [7] J. Friedman, T. Hastie, and R. Tibshirani. Additive logistic regression: A statistical view of boosting. The Annals of Statistics, 28(2), 2000. [8] J. Kivinen and M. K. Warmuth. Boosting as entropy projection. In Computational Learning Theory, 1999. [9] J. Lafferty. Additive models, boosting, and inference for generalized divergences. In Computational Learning Theory, 1999. [10] L. Mason, J. Baxter, P. Bartlett, and M. Frean. Functional gradient techniques for combining hypotheses. In A. Smola, P. Bartlett, B. Sch?olkopf, and D. Schuurmans, editors, Advances in Large Margin Classifiers, 1999. [11] G. R?atsch, T. Onoda, and K.-R. M?uller. Soft margins for AdaBoost. Machine Learning, 2001.
2042 |@word version:1 pw:3 closure:1 incurs:1 moment:1 com:1 must:1 john:1 additive:2 enables:1 plot:1 interpretable:1 update:3 v_:7 discrimination:1 intelligence:1 selected:1 yr:12 warmuth:2 boosting:34 direct:1 become:1 psfrag:6 vad:2 introduce:1 expected:2 multi:1 little:1 increasing:1 becomes:4 ua:3 moreover:2 notation:2 v7t:2 mass:1 qw:1 what:1 interpreted:1 unified:1 unobserved:1 berkeley:1 every:1 returning:1 classifier:2 appear:1 understood:1 establishing:1 au:5 equivalence:1 suggests:2 gone:1 averaged:2 practical:1 unique:2 hood:1 yj:1 testing:1 acknowledgment:1 differs:2 procedure:6 empirical:4 projection:2 spite:1 suggest:1 get:3 onto:2 convenience:1 context:2 writing:1 equivalent:2 lagrangian:2 straightforward:1 convex:7 survey:1 helmbold:2 m2:5 insight:2 rule:1 mq:1 annals:1 suppose:2 hypothesis:1 pa:2 trend:1 satisfying:3 labeled:2 bottom:3 calculate:1 removed:1 trained:3 likeli:1 solving:1 learner:4 easily:1 various:2 intersects:2 train:19 whose:1 emerged:1 loglikelihood:1 statistic:6 rosenfeld:1 noisy:1 unresolved:1 uci:3 combining:1 flexibility:1 inducing:1 olkopf:1 qr:4 convergence:1 empty:1 produce:1 derive:8 andrew:1 n_:1 frean:1 measured:2 rescale:1 school:2 dividing:1 auxiliary:1 c:3 indicate:2 come:1 convention:1 kuhn:1 direction:1 closely:2 require:1 proposition:1 pl:2 c_:1 hold:5 cb:1 prm:1 lm:1 early:1 omitted:1 purpose:2 estimation:2 f7:2 label:2 e45:1 uller:1 gaussian:2 e7:1 rather:3 avoid:1 sion:1 notational:1 consistently:2 likelihood:35 mainly:1 hepatitis:2 hk:1 sense:1 glass:2 helpful:1 inference:1 kkw:2 arg:1 dual:16 denoted:2 constrained:1 special:4 smoothing:1 uc:1 marginal:1 field:1 once:2 ng:1 identical:3 simplex:2 minimized:1 report:1 employ:1 randomly:2 divergence:12 resulted:1 pietra:4 familiar:1 geometry:1 consisting:1 replacement:6 friedman:2 interest:3 investigate:1 primal:4 devoted:1 kt:1 bregman:7 necessary:1 old:1 harmful:1 circle:1 theoretical:3 increased:2 column:3 modeling:1 soft:1 cost:1 entry:2 uniform:1 gcv:1 considerably:1 international:1 e_:1 diverge:1 michael:2 connecting:2 together:1 satisfied:1 possibly:1 guy:1 booster:1 jrk:1 account:2 potential:3 stump:1 satisfy:2 reg:4 performed:2 view:6 parallel:2 identifiability:1 minimize:5 square:1 correspond:4 yield:1 weak:4 bayesian:2 produced:1 tucker:1 proof:1 irvine:1 proved:1 dataset:4 improves:4 appears:2 higher:3 adaboost:19 wei:1 unidentifiable:1 generality:1 smola:1 until:1 overfit:1 working:1 lent:1 hand:1 nonlinear:2 lack:2 logistic:11 grows:2 normalized:8 multiplier:5 consisted:1 regularization:10 equality:1 leibler:3 ll:1 noted:1 unnormalized:4 iris:2 generalized:1 mpl:5 demonstrate:1 neo:1 recently:1 functional:2 empirically:1 exponentially:1 interpretation:2 mellon:3 refer:1 significant:1 trivially:1 language:1 similarity:1 add:3 xwzy:1 recent:2 xzw:1 binary:4 sjt:1 seen:3 minimum:1 additional:6 impose:1 determine:3 maximize:1 fernando:1 violate:1 reduces:1 technical:1 match:1 calculation:1 cross:2 raphson:1 e1:1 uav:2 regression:10 basic:1 breast:1 cmu:3 iteration:1 represent:1 normalization:4 addition:2 affecting:1 leaving:1 sch:1 zw:8 qrw:1 comment:1 subject:4 member:1 lafferty:6 n7:1 effectiveness:1 jordan:1 easy:2 relaxes:1 baxter:1 fit:3 hastie:2 bartlett:2 k9:1 penalty:1 speech:1 v8:1 concentrated:1 schapire:4 exist:1 problematic:1 shifted:1 tibshirani:2 carnegie:3 key:2 falling:1 clean:1 striking:1 throughout:2 family:4 decision:1 def:2 correspondence:3 fold:2 quadratic:3 nonnegative:1 activity:1 precisely:2 constraint:21 attempting:1 department:1 according:1 conjugate:1 smaller:3 y0:3 rob:1 making:2 happens:1 projecting:2 unregularized:3 equation:4 agree:1 singer:2 b_:9 appropriate:1 fad:9 alternative:2 yair:1 original:1 denotes:1 top:3 cf:1 newton:1 giving:1 especially:1 nboost:2 objective:2 usual:2 visiting:1 gradient:2 distance:7 reversed:1 thank:1 me:1 relationship:8 minimizing:5 setup:1 statement:1 pima:1 expense:1 negative:1 ba:1 datasets:5 finite:2 descent:1 extended:4 excluding:1 precise:2 dc:1 arbitrary:1 required:2 specified:1 kl:4 connection:5 california:1 boost:26 pattern:1 including:1 explanation:1 natural:1 regularized:4 kivinen:2 mn:2 carried:2 kj:2 prior:4 understanding:1 geometric:2 freund:1 loss:21 validation:2 sufficient:3 consistent:3 ponential:1 editor:1 lmn:3 row:5 course:1 penalized:2 ovum:1 formal:1 side:1 understand:1 taking:2 default:1 calculated:1 doesn:1 kau:2 forward:1 qualitatively:1 author:1 transaction:2 lebanon:2 approximate:1 kullback:3 ml:36 suggestive:1 overfitting:2 pittsburgh:2 sonar:3 table:3 fao:2 k_:1 zk:1 onoda:1 f_:12 schuurmans:1 da:1 mnl:14 promoter:2 linearly:1 tl:1 fails:1 lmo:1 pereira:1 exponential:29 lie:1 ib:1 third:1 theorem:5 showing:1 pac:1 mason:2 ionosphere:1 a3:1 normalizing:3 essential:1 stepwise:2 sequential:2 duffy:2 margin:3 chen:1 entropy:4 intersection:1 simply:1 lagrange:5 corresponds:7 conditional:12 viewed:2 formulated:1 presentation:3 feasible:4 specifically:1 total:1 duality:8 e:1 experimental:1 atsch:1 formally:1 support:2 latter:3 collins:3 overload:1 pythagorean:1 audio:1 regularizing:1 della:4 ex:1
1,144
2,043
Efficient Resources Allocation for Markov Decision Processes Remi Munos CMAP, Ecole Polytechnique, 91128 Palaiseau, France http://www.cmap.polytechnique.fr/....munos remi.munos@polytechnique.fr Abstract It is desirable that a complex decision-making problem in an uncertain world be adequately modeled by a Markov Decision Process (MDP) whose structural representation is adaptively designed by a parsimonious resources allocation process. Resources include time and cost of exploration, amount of memory and computational time allowed for the policy or value function representation. Concerned about making the best use of the available resources, we address the problem of efficiently estimating where adding extra resources is highly needed in order to improve the expected performance of the resulting policy. Possible application in reinforcement learning (RL) , when real-world exploration is highly costly, concerns the detection of those areas of the state-space that need primarily to be explored in order to improve the policy. Another application concerns approximation of continuous state-space stochastic control problems using adaptive discretization techniques for which highly efficient grid points allocation is mandatory to survive high dimensionality. Maybe surprisingly these two problems can be formulated under a common framework: for a given resource allocation, which defines a belief state over possible MDPs, find where adding new resources (thus decreasing the uncertainty of some parameters -transition probabilities or rewards) will most likely increase the expected performance of the new policy. To do so, we use sampling techniques for estimating the contribution of each parameter's probability distribution function (Pdf) to the expected loss of using an approximate policy (such as the optimal policy of the most probable MDP) instead of the true (but unknown) policy. Introduction Assume that we model a complex decision-making problem under uncertainty by a finite MDP. Because of the limited resources used, the parameters of the MDP (transition probabilities and rewards) are uncertain: we assume that we only know a belief state over their possible values. IT we select the most probable values of the parameters, we can build a MDP and solve it to deduce the corresponding optimal policy. However, because of the uncertainty over the true parameters, this policy may not be the one that maximizes the expected cumulative rewards of the true (but partially unknown) decision-making problem. We can nevertheless use sampling techniques to estimate the expected loss of using this policy. Furthermore, if we assume independence of the parameters (considered as random variables), we are able to derive the contribution of the uncertainty over each parameter to this expected loss. As a consequence, we can predict where adding new resoUrces (thus decreasing the uncertainty over some parameters) will decrease mostly this loss, thus improving the MDP model of the decision-making problem so as to optimize the expected future rewards. As possible application, in model-free RL we may wish to minimize the amount of real-world exploration (because each experiment is highly costly). Following [1] we can maintain a Dirichlet pdf over the transition probabilities of the corresponding MDP. Then, our algorithm is able to predict in which parts of the state space we should make new experiments, thus decreasing the uncertainty over some parameters (the posterior distribution being less uncertain than the prior) in order to optimize the expected payoff. Another application concerns the approximation of continuous (or large discrete) state-space control problems using variable resolution grids, that requires an efficient resource allocation process in order to survive the "curse of dimensionality" in high dimensions. For a given grid, because of the interpolation process, the approximate back-up operator introduces a local interpolation error (see [4]) that may be considered as a random variable (for example in the random grids of [6]). The algorithm introduced in this paper allows to estimate where we should add new grid-points, thus decreasing the uncertainty over the local interpolation error, in order to increase the expected performance of the new grid representation. The main tool developed here is the calculation of the partial derivative of useful global measures (the value function or the loss of using a sub-optimal policy) with respect to each parameter (probabilities and rewards) of a MDP. 1 Description of the problem We consider a MDP with a finite state-space X and action-space A. A transition from a state x, action a to a next state y occurs with probability p(Ylx, a) and the corresponding (deterministic) reward is r(x, a). We introduce the back-up operator T a defined, for any function W : X --t JR, as T a W(x) == (' LP(Ylx, a)W(y) + r(x, a) (1) y (with some discount factor 0 < (' < 1). It is a contraction mapping, thus the dynamic programming (DP) equation V(x) == maxaEA T a V(x) has a unique fixed point V called the value function. Let. us define the Q-values Q(x, a) == T a V (x). The optimal policy 1[* is the mapping from any state x to? the action 1[* (x) that maximizes the Q-values: 1[*(x) == maxaEA Q(x, a). The parameters of the MDP - the probability and the reward functions - are not perfectly known: all we know is a pdf over their possible values. This uncertainty comes from the limited amount of allocated resources for estimating those parameters. Let us choose a specific policy 1r (for example the optimal policy of the MDP with the most probable parameters). We can estimate the expected loss of using 1r instead of the true (but unknown) optimal policy 1[*. Let us write J-t == {Pj} the set of all parameters (p and r functions) of a MDP. We assume that we know a probability distribution function pdf(J-Lj) over their possible values. For a MDP MJ.t defined by its parameters P, we write pJL (y Ix, a), r JL (x, a), V JL, QJL, and 7f1-!' respectively its transition probabilities, rewards, value function, Q-values, and optimal policy. 1.1 Direct gain optimization We define the gain ]JL(x; 7f) in the MDP MJL as the expected sum of discounted rewards obtained starting from state x and using policy 7f: ]JL(x; 1f) == E[2: rykrJL(Xk' 7f(xk))lxo == x; 7f] (2) k where the expectation is taken for sequences of states Xk --t Xk+l occurring with probability pP(Xk+llxk, 7fJL(Xk)). By definition, the optimal gain in MJL is VJL(x) == ]JL (x; 7fJL) which is obtained for the optimal policy 7fIL. Let ~ (x) == ]JL (x; if) be the approximate gain obtained for some approximate policy .7r in the same MDP MIL. We define the loss to occur LJL(x) from X when one uses the approximate policy 7r instead of the optimal one 7fJL in MJL: LIL(X) == VIL(x) - ~(x) (3) An example of approximate policy 1? would be the optimal policy of the most probable MDP, defined by the most probable parameters fi(ylx, a) and r(x, a). We also consider the problem of maximizing the global gain from a set of initial states chosen according to some probability distribution P(x). Accordingly, we define the global gain of a policy 11"": ]JL(7f) == Ex ]JL(x; 7f)P(x) and the global loss LIL of using some approximate policy 7r instead of the optimal one nIL (4) Thus, knowing the pdf over all parameters J-l we can define the expected global loss L == EJL[LIL]. Next, we would like to define what is the contribution of each parameter uncertainty to this loss, so we know where we should add new resources (thus reducing some parameters uncertainty) in order to decrease the expected global loss. We would like to estimate, for each parameter J-lj, (5) E[8L I Add 8u units of resource for Pj] 1.2 Partial derivative of the loss ill order to quantify (5) we need to be more explicit about the pdf over JL. First, we assume the independence of the parameters JLj (considered as random variables). Suppose that pdf (JLj) == N (0, U j) (normal distribution of mean 0 and standard deviation Uj). We would like to estimate the variation 8L of the expected loss L when we make a small change of the uncertainty over Pj (consequence of adding new resources), for example when changing the standard deviation of 8aj in pdf(J.tj). At the limit of an infinitesimal variation we obtain the partial derivative which 3 when computed for all parameters J-lj, provides the respective contributions of each parameter's uncertainty to the global loss. Z;., Another example is when the pdf(pj) is a uniform distribution of support [-b j , bj ]. Then the partial contribution of JLj'S uncertainty to the global loss can be expressed More generally, we can define a finite number of characteristic scalar meaas 3 surements of the pdf uncertainty (for example the entropy or the moments) and gf. compute the partial derivative of the expected global loss with respect to these coefficients. Finally, knowing the actual resources needed to estimate a parameter J..tj with some uncertainty defined by pdf (J..tj ), we are able to estimate (5). 1.3 Unbiased estimator We sample N sets of parameters {J..t i }i=1..N from the pd!(J..t) , which define N-MDPs Mi. For convenience, we use the superscript i to refer to the i-th MDP sample and the subscript j for the j-th parameter of a variable. We solve each MDP using standard DP techniques (see [5]). This expensive computation that can be speed-up in two ways: first, by using the value function and policy computed for the first MDP as initial values for the other MDPs; second, since all MDPs have the same structure, by computing once for all an efficient ordering (using a topological sort, possibly with loops) of the states that will be used for value iteration. For each MDP, we compute the global loss L i of using the policy 'if and estimate the expected global loss: L ~ 2:::1 L i . In order to estimate the contribution of a p-arameter's uncertainty to L, we derive the partial derivative of L with respect to the characteristic coefficients of pdf (J-tj ). In the case of a reward parameter J..tj that follows a normal distribution N(O, Uj), we can write J..tj == Uj?j where ?j follows N(O, 1). The partial derivative of the expected loss L with respect to Uj is -1 8 8 aL == a E/L~N(o.u)[L/L] = a Ee~N(o.l)[LUe] = a~ ~ ~ Ee~N(o.1)[8aLue ~j] (6) ~ from which we deduce the unbiased estimator 8L '" ~ aUj - N t i=l i Jt; (7) 8L aJ..tj Uj where ~;; is the partial derivative of the global loss Li of MDP M i with respect to the parameter J..tj (considered as a variable). For other distributions, we can define similar results to (6) and deduce analogous estimators (for uniform distributions, we have the same estimator with bj instead of Uj). The remainder of the paper is organized as follow. Section 2 introduces useful tools to derive the partial contribution of each parameter -transition probability and reward- to the value function in a Markov Chain, Section 3 establishes the partial contribution of each parameter to the global loss, allowing to calculate the estimator (7), and Section 4 provides an efficient algorithm. All proofs are given in the full length paper [2]. 2 2.1 Non-local dependencies Influence of a markov chain In [3] we introduced the notion of influence of a Markov Chain as a way to measure value function/rewards correlations between states. Let us consider a set of values V satisfying a Bellman equation Vex) == , LP(ylx)V(y) + rex) (8) y We define the discounted cumulative k-chained transition probabilities Pk(ylx): po(ylx) Pl(ylx) (= 1 (if x = y) or 0 (if x IP(ylx) Ix =y =1= y)) LP1(ylw)Pl(wlx) w LP1(ylw)Pk-l(wlx) w The influence I(ylx) of a state y on another state x is defined as I(ylx) = 2::%:oPk(ylx). Intuitively, I(ylx) measures the expected discounted number of visits of state y starting from x; it is also the partial derivative of the value function Vex) with respect to the reward r(y). Indeed Vex) can be expressed, as a linear combination of the rewards at y weighted by the influence I(ylx) (9) Vex) = LI(Ylx)r(y) y We can also define the influence of a state y on a function f: I(ylf(?)) = 2::x l(ylx)f(x) and the influence of a function f on another function 9 : l(f(?)\g(?)) = Y":y I(ylg(?))f(y)? In [3], we showed that the influence satisfies I(ylx) =, LP(ylw)I(wlx) + lx=y (10) w which is a fixed-point equation of a contractant operator (in I-norm) thus has a unique solution -the influence- that can be computed by successive iterations. Similarly, the influence I(ylf(?)) can be obtained as limit of the iterations I(ylf(?)) +-, LP(Ylw)I(wlf(?)) + fey) w Thus the computation of the influence I(ylf(?)) is cheap (equivalent to solving a Markov chain). 2.2 Total derivative of V We wish to express the contribution of all parameters - transition probabilities p and rewards r - (considered as variables) to the value function V by defining the total derivative of V as a function of those P?ameters. We recall that the total f dXI + ... + a8t dx . derivative of a function f of several variables Xl, ..,' X n is df = 88Xl n Xn We already know that the partial derivative of Vex) with respect to the reward r(z) is the influence I(zjx) = ~~~1. Now, the dependency with respect to the transition probabilities has to be expressed more carefully because the probabilities p(wlz) for a given z are dependent (they sum to one). A way to express that is provided in the theorem that follows whose proof is in [2]. Theorelll 1 For a given state z, let us alter the probabilities p(wlz), for all w, with some c5'p(wlz) value, such that 2:: w c5'p(wlz) = o. Then Vex) is altered by c5'V(x) = I(zlx)[,2:: w V(w)c5'p(wlz)]. We deduce the total derivative of v: dV(x) = L1(zlx)[, L z V(w)dp(wlz) + dr(z)] w under the constraint 2::w dp( wi z) = 0 for all z. 3 Total derivative of the loss , For a given MDP M with parameters J..L (for notation simplification we do not write the JL superscript in what follows), we want to estimate the loss of using an approximate policy 7? instead of the optimal one 1f. First, we define the one-step loss l(x) at a state x as the difference between the gain obtained by choosing the best action 7f(x) then using the optimal policy 1f and the gain obtained by choosing action n(x) then the same optimal policy 7f l(x) == Q(x,1f(x)) - Q(x,ir(x)) (11) Now we consider the loss L(x), defined by (3), for an initial state x when we use the approximate policy n. We can prove that L(x) is the expected discounted cumulative one-step losses l(Xk) for reachable states Xk: L(x) == E[L I'k l(Xk)lxo == x;n] k with the expectation taken in the same sense as in (2). 3.1 Decomposition of the one-step loss We use (9) to decompose the Q-values Q(x, a) == I' LP(wlx, a) L I(ylw)r(y, 1f(Y)) w + r(x, a) y == r(x,a) + Lq(Ylx,a)r(y,7f(y)) y using the partial contributions q(ylx,a) == I'Ewp(wlx,a)I(ylw) where I(ylw) is the influence of y on w in the Markov chain derived from the MDP M by choosing policy 7f. Similarly, we decompose the one-step loss l(x) == Q(x,7f(x)) - Q(x, n(x)) == r(x,1f(x)) - r(x,7f(x)) + L [q(ylx,1f(x)) - q(ylx,n(x))] r(y,7f(Y)) y == r(x, 7f(x)) -r(x, 7?(x)) + Ll(Ylx)r(y, 7f(Y)) y as a function of the partial contributions l(ylx) == q(ylx,1f(x)) - q(ylx, n(x)) (see figure 1). o q (ylx ,IT ) q (ylx ,11- ) Figure 1: The reward r(y,1r(Y)) at state y contributes to the one-step loss l(x) = Q(x, 1r(x)) - Q(x, 1?(x)) with the proportion l(ylx) q(ylx, 1I"(x)) - q(ylx, 1?(x)). 3.2 Total derivative of the one-step loss and global loss Similarly to section (2.2), we wish to express the contribution of all parameters transition probabilities p and rewards r - (considered as variables) to the one-step loss function by defining the total derivative of I as a function of those parameters. Theorem 2 Let us introduce the (formal) differential back-up operator dT a defined, for any function W : X ~ JR, as dT a W(x) == ry L W(y)dp(ylx, a) + dr(x, a) y (similar to the back-up operator (1) but using dp and dr instead of p and r). The total derivative of the one-step loss is dl(x).==L 1(zlx)dT 7f (z)V(z) + dT7f(x)V(x) - dT;Cx)V(x) z under the constraint E y dp(ylx, a) == 0 for all x and a. Theorem 3 Let us introduce the one-step-loss back-up operator S and its formal differential version dS defined, for any function W : X ~ JR, as SW(x) ry LP(Ylx, 7T"(x))W(y) + l(x) y dSW(x) ry L dp(ylx, 7T"(x))W(y) + dl(x) y Then, the loss L(x) at x satisfies Bellman's equation L of the loss L (x) and global loss L are, respectively dL(x) == SL. The total derivative L I(zlx)dSL(z) Z dL L I(zIP(?))dSL(z) z from which (after regrouping the contribution to each parameter) we deduce the partial derivatives of the global loss with respect to the rewards and transition probabilities 4 A fast algorithm We use the sampling technique introduced in section 1.3. In order to compute the estimator (7) we calculate the partial derivatives ~~; based on the result of the previous section, with the following algorithm~ Given the pdf over the parameters j.L, select a policy 7? (for example the optimal policy of the most probable MDP). For i == 1..N, solve each MDP M i and deduce its value function Vi, Q-values Qi, and optimal policy 7ri . Deduce the one-step loss li(x) from (11). Compute the influence I(xIP(?)) (which depends on the transition probabilities pi of M i ) and the influence I(li(xl?)IP(?)) from which we deduce i ar ~(Lix,a ). Then calculate Li(x) by solving Bellman's equation Li = SL and deduce 8P,r~:,a). These partial derivatives enable to compute the unbiased estimator (7). The complexity of solving a discounted MDP with K states, each one connected to M next states, is O(KM), as is the complexity of computing the influences. Thus, the overall complexity of this algorithm is O(NKM). Conclusion? Being able to compute the contribution of each parameter -transition probabilities and rewards- to the value function (theorem 1) and to the loss of the expected rewards to occur if we use an approximate policy (theorem 3) enables us to use sampling techniques to estimate what are the parameters whose uncertainty are the most harmful to the expected gain.. A relev-ant resource allocation process would consider adding new computational resources to reduce uncertainty over the true value of those parameters. In the examples given in the introduction, this would be doing new experiments in model-free RL for defining more precisely the transition probabilities of some relevant states. In discretization techniques for continuous control problems, this would be adding new grid points in order to improve the quality of the interpolation at relevant areas of the state space in order to maximize the expected gain of the new policy. Initial experiments for variable resolution discretization using random grids show improved performance compared to [3]. Acknowledgments I am grateful to Andrew Moore, Drew Bagnell and Auton's Lab members for motivating discussions. References [1] Richard Dearden, Nir Friedman, and David Andre. Model based bayesian exploration. Proceeding of Uncertainty in Artificial Intelligence, 1999. [2] Remi Munos. Decision-making under uncertainty:. Efficiently estimating where extra ressources are needed. Technical report, Ecole Polytechnique, 2002. [3] Remi Munos and Andrew Moore. Influence and variance of a markov chain : Application to adaptive discretizations in optimal control. Proceedings of the 38th IEEE Conference on Decision and Control, 1999. [4] Remi Munos and Andrew W. Moore. Rates of convergence for variable resolution schemes in optimal control. International Conference on Machine Learning, 2000. [5] Martin L. Puterman. Markov Decision Processes, Discrete Stochastic Dynamic Programming. A Wiley-Interscience Publication, 1994. [6] John Rust. Using Randomization to Break the Curse of Dimensionality. Computational Economics. 1997.
2043 |@word version:1 norm:1 proportion:1 km:1 contraction:1 decomposition:1 moment:1 initial:4 ecole:2 discretization:3 dx:1 john:1 cheap:1 enables:1 lue:1 designed:1 intelligence:1 accordingly:1 xk:9 provides:2 lx:1 successive:1 direct:1 differential:2 prove:1 interscience:1 ressources:1 introduce:3 indeed:1 expected:22 ry:3 bellman:3 discounted:5 decreasing:4 actual:1 curse:2 provided:1 estimating:4 notation:1 maximizes:2 what:3 developed:1 control:6 unit:1 local:3 limit:2 consequence:2 subscript:1 interpolation:4 limited:2 unique:2 acknowledgment:1 area:2 discretizations:1 convenience:1 operator:6 influence:16 www:1 optimize:2 deterministic:1 equivalent:1 maximizing:1 economics:1 starting:2 resolution:3 estimator:7 notion:1 variation:2 analogous:1 suppose:1 programming:2 us:1 jlj:3 expensive:1 satisfying:1 lix:1 calculate:3 connected:1 ordering:1 decrease:2 pd:1 complexity:3 reward:21 dynamic:2 chained:1 grateful:1 solving:3 po:1 fast:1 artificial:1 choosing:3 whose:3 solve:3 superscript:2 ip:2 sequence:1 fr:2 remainder:1 relevant:2 loop:1 description:1 convergence:1 derive:3 andrew:3 come:1 quantify:1 stochastic:2 exploration:4 enable:1 alue:1 f1:1 decompose:2 theorelll:1 randomization:1 probable:6 auj:1 pl:2 fil:1 considered:6 normal:2 mapping:2 predict:2 bj:2 establishes:1 tool:2 weighted:1 nkm:1 mil:1 publication:1 derived:1 sense:1 am:1 dependent:1 xip:1 lj:3 france:1 overall:1 ill:1 once:1 sampling:4 survive:2 a8t:1 future:1 alter:1 report:1 richard:1 primarily:1 maintain:1 friedman:1 detection:1 highly:4 wlf:1 introduces:2 tj:8 chain:6 partial:17 respective:1 harmful:1 uncertain:3 ar:1 opk:1 cost:1 deviation:2 uniform:2 rex:1 motivating:1 dependency:2 adaptively:1 international:1 choose:1 possibly:1 dr:3 derivative:21 li:6 zjx:1 coefficient:2 vi:1 depends:1 break:1 lab:1 doing:1 surements:1 sort:1 contribution:14 minimize:1 ir:1 variance:1 characteristic:2 efficiently:2 ant:1 dsl:2 bayesian:1 vil:1 andre:1 definition:1 infinitesimal:1 pp:1 proof:2 mi:1 dxi:1 gain:10 recall:1 dimensionality:3 organized:1 carefully:1 back:5 dt:4 follow:1 improved:1 furthermore:1 correlation:1 d:1 defines:1 quality:1 aj:2 mdp:26 fjl:3 true:5 unbiased:3 wlz:6 adequately:1 moore:3 puterman:1 ll:1 pdf:13 polytechnique:4 l1:1 fi:1 common:1 rl:3 rust:1 jl:10 ejl:1 refer:1 grid:8 similarly:3 reachable:1 ylf:4 deduce:9 add:3 posterior:1 showed:1 mandatory:1 regrouping:1 zip:1 maximize:1 full:1 desirable:1 technical:1 calculation:1 lxo:2 vjl:1 visit:1 qi:1 expectation:2 df:1 iteration:3 want:1 allocated:1 extra:2 member:1 structural:1 ee:2 concerned:1 independence:2 perfectly:1 reduce:1 knowing:2 action:5 useful:2 generally:1 llxk:1 ylx:33 maybe:1 amount:3 discount:1 http:1 ylw:7 sl:2 qjl:1 discrete:2 write:4 express:3 nevertheless:1 changing:1 pj:4 sum:2 uncertainty:20 parsimonious:1 vex:6 decision:9 simplification:1 topological:1 occur:2 constraint:2 precisely:1 ri:1 speed:1 martin:1 according:1 combination:1 jr:3 fey:1 ewp:1 wi:1 lp:6 making:6 arameter:1 intuitively:1 dv:1 taken:2 resource:17 equation:5 needed:3 know:5 auton:1 available:1 dirichlet:1 include:1 sw:1 build:1 uj:6 already:1 occurs:1 costly:2 bagnell:1 dp:8 length:1 modeled:1 mostly:1 lil:3 policy:36 unknown:3 allowing:1 markov:9 finite:3 payoff:1 defining:3 introduced:3 david:1 zlx:4 address:1 able:4 maxaea:2 memory:1 belief:2 dearden:1 scheme:1 improve:3 altered:1 mdps:4 gf:1 nir:1 prior:1 loss:40 allocation:6 pi:1 surprisingly:1 free:2 formal:2 munos:6 dimension:1 xn:1 world:3 transition:14 cumulative:3 c5:4 reinforcement:1 adaptive:2 palaiseau:1 approximate:10 global:16 continuous:3 ljl:1 mj:1 pjl:1 contributes:1 improving:1 complex:2 pk:2 main:1 allowed:1 wiley:1 sub:1 wish:3 explicit:1 lq:1 xl:3 ix:2 theorem:5 specific:1 jt:1 explored:1 concern:3 dl:4 adding:6 drew:1 cmap:2 occurring:1 entropy:1 cx:1 remi:5 likely:1 expressed:3 partially:1 scalar:1 satisfies:2 formulated:1 change:1 reducing:1 called:1 nil:1 total:9 lp1:2 select:2 support:1 ex:1
1,145
2,044
Speech Recognition using SVMs Nathan Smith Cambridge University Engineering Dept Cambridge, CB2 1PZ, U.K. ndsl 002@eng.cam.ac.uk Mark Gales Cambridge University Engineering Dept Cambridge, CB2 1PZ, U.K. mjfg@eng.cam.ac.uk Abstract An important issue in applying SVMs to speech recognition is the ability to classify variable length sequences. This paper presents extensions to a standard scheme for handling this variable length data, the Fisher score. A more useful mapping is introduced based on the likelihood-ratio. The score-space defined by this mapping avoids some limitations of the Fisher score. Class-conditional generative models are directly incorporated into the definition of the score-space. The mapping, and appropriate normalisation schemes, are evaluated on a speaker-independent isolated letter task where the new mapping outperforms both the Fisher score and HMMs trained to maximise likelihood. 1 Introduction Speech recognition is a complex, dynamic classification task. State-of-the-art systems use Hidden Markov Models (HMMs), either trained to maximise likelihood or discriminatively, to achieve good levels of performance. One of the reasons for the popularity of HMMs is that they readily handle the variable length data sequences which result from variations in word sequence, speaker rate and accent. Support Vector Machines (SVMs) [1] are a powerful, discriminatively-trained technique that have been shown to work well on a variety of tasks. However they are typically only applied to static binary classification tasks. This paper examines the application of SVMs to speech recognition. There are two major problems to address. First, how to handle the variable length sequences. Second, how to handle multi-class decisions. This paper only concentrates on dealing with variable length sequences. It develops our earlier research in [2] and is detailed more fully in [7]. A similar approach for protein classification is adopted in [3]. There have been a variety of methods proposed to map variable length sequences to vectors of fixed dimension. These include vector averaging and selecting a 'representative ' number of observations from each utterance. However, these methods may discard useful information. This paper adopts an approach similar to that of [4] which makes use of all the available data. Their scheme uses generative probability models of the data to define a mapping into a fixed dimension space, the Fisher score-space. When incorporated within an SVM kernel, the kernel is known as the Fisher kernel. Relevant regularisation issues are discussed in [5]. This paper examines the suitability of the Fisher kernel for classification in speech recognition and proposes an alternative, more useful, kernel. In addition some normalisation issues associated with using this kernel for speech recognition are addressed. Initially a general framework for defining alternative score-spaces is required. First, define an observation sequence as 0 = (01 , . . . Ot, ... OT) where Ot E ~D , and a set of generative probability models of the observation sequences as P = {Pk(OI(h)}, where 9 k is the vector of parameters for the kth member of the set. The observation sequence 0 can be mapped into a vector of fixed dimension [4], i{J~ (0) (1) f(?) is the score-argument and is a function of the members of the set of generative models P. i{Jft is the score-mapping and is defined using a score-operator F. i{J~(0) is the score and occupies the fixed-dimension score-space. Our investigation of score-spaces falls into three divisions. What are the best generative models, scorearguments and score-operators to use? 2 Score-spaces As HMMs have proved successful in speech recognition, they are a natural choice as the generative models for this task. In particular HMMs with state output distributions formed by Gaussian mixture models. There is also the choice of the score-argument. For a two-class problem, let Pi(019 i ) represent a generative model, where i = {g, 1, 2} (g denotes the global2-class generative model, and 1 and 2 denote the class-conditional generative models for the two competing classes). Previous schemes have used the log of a single generative model, Inpi (019 i ) representing either both classes as in the original Fisher score (i = g) [4], or one of the classes (i = 1 or 2) [6]. This score-space is termed the likelihood score-space, i{J~k(O). The score-space proposed in this paper uses the log of the ratio of the two classconditional generative models, In(P1(019d / P2(019 2)) where 9 = [9{,9J] T. The corresponding score-space is called the likelihood-ratio score-space, i{J~(0) . Thus, i{J~k(O) (2) i{J~(0) (3) The likelihood-ratio score-space can be shown to avoid some of the limitations of the likelihood score-space, and may be viewed as a generalisation of the standard generative model classifier. These issues will be discussed later. Having proposed forms for the generative models and score-arguments, the scoreoperators must be selected. The original score-operator in [4] was the 1st-order derivative operator applied to HMMs with discrete output distributions. Consider a continuous density HMM with N emitting states, j E {I . . . N}. Each state, j, has an output distribution formed by a mixture of K Gaussian components, N(J-tjk' ~jd where k E {I ... K}. Each component has parameters of weight Wjk, mean J-tjk and covariance ~jk. The 1st-order derivatives of the log probability of the sequence 0 with respect to the model parameters are given below1, where the derivative operator has been defined to give column vectors, T L ')'jk(t)S~,jkl t= l lFor fuller details of the derivations see [2). (4) V Wjk Inp(OIO) where S[t ,jk] Ijdt) is the posterior probability of component k of state j at time t. Assuming the HMM is left-to-right with no skips and assuming that a state only appears once in the HMM (i.e. there is no state-tying), then the 1st-order derivative for the self-transition probability for state j, ajj, is, t[/j(t) t=l ajj 1] Tajj(l- ajj) (8) The 1st-order derivatives for each Gaussian parameter and self-transition probability in the HMM can be spliced together into a 'super-vector' which is the score 2 . From the definitions above, the score for an utterance is a weighted sum of scores for individual observations. If the scores for the same utterance spoken at different speaking rates were calculated, they would lie in different regions of score-space simply because of differing numbers of observations. To ease the task of the classifier in score-space, the score-space may be normalised by the number of observations, called sequence length normalisation. Duration information can be retained in the derivatives of the transition probabilities. One method of normalisation redefines score-spaces using generative models trained to maximise a modified log likelihood function, In( 010). Consider that state j has entry time Tj and duration d j (both in numbers of observations) and output probability bj(Ot) for observation Ot [7]. So, 1 N L In(OIO) T;+d j- 1 d- ((d j -1) lnajj + Inaj(j+1) + j=l L (Inbj(Ot))) (9) t=Tj J It is not possible to maximise In(OIO) using the EM algorithm. Hill-climbing techniques could be used. However, in this paper, a simpler normalisation method is employed. The generative models are trained to maximise the standard likelihood function. Rather than define the score-space using standard state posteriors Ij(t) (the posterior probability of state j at time t), it is defined on state posteriors normalised by the total state occupancy over the utterance. The standard component posteriors 1 j k (t) are replaced in Equations 4 to 6 and 8 by their normalised form 'Yjk(t), A . ~k (t) _ - Ij(t) T (WjkN(Ot; ILjk, ~jk) K 2:: T=l/j(T) 2:: i = l wjiN(ot; ILji' ~ji) ) (10) In effect, each derivative is divided by the sum of state posteriors. This is preferred to division by the total number of observations T which assumes that when the utterance length varies, the occupation of every state in the state sequence is scaled by the same ratio. This is not necessarily the case for speech. The nature of the score-space affects the discriminative power of classifiers built in the score-space. For example, the likelihood score-space defined on a two-class 2Due to the sum to unity constraints, one of the weight parameters in each Gaussian mixture is discarded from the definition of the super-vector, as are the forward transitions in the left-to-right HMM with no skips. generative model is susceptible to wrap-around [7] . This occurs when two different locations in acoustic-space map to a single point in score-subspace. As an example, consider two classes modelled by two widely-spaced Gaussians. If an observation is generated at the peak of the first Gaussian, then the derivative relative to the mean of that Gaussian is zero because S [t ,jk] is zero (see Equation 4). However, the derivative relative to the mean of the distant second Gaussian is also zero because of a zero component posterior f jdt). A similar problem occurs with an observation generated at the peak of the second Gaussian. This ambiguity in mapping two possible locations in acoustic-space to the zero of the score-subspace of the means represents a wrapping of the acoustic space onto this subspace. This also occurs in the subspace of the variances. Thus wrap-around can increase class confusion. A likelihood-ratio score-space defined on these two Gaussians does not suffer wraparound since the component posteriors for each Gaussian are forced to unity. So far, only 1st-order derivative score-operators have been considered. It is possible to include the zeroth, 2nd and higher-order derivatives. Of course there is an interaction between the score-operator and the score-argument. For example, the zeroth-order derivative for the likelihood score-space is expected to be less useful than its counter-part in the likelihood-ratio score-space because of its greater sensitivity to acoustic conditions. A principled approach to using derivatives in score-spaces would be useful. Consider the simple case of true class-conditional generative models P1(OIOd and P2(OI02) with respective estimates of the same functional form P1 (0 10d and P2(010 2 ) . Expressing the true models as Taylor series expansions about the parameter estimates 01 and O2 (see [7] for more details, and [3]) , Inpi (OIOi ) + (Oi - Oi ) TV' 9i Inpi (OIOi ) 1 A T T A A ( +"2(Oi - Oi ) [V' 9i V' 9i Inpi (OIOi )](Oi - Oi ) + 0 Oi (?) will , V'~i' vec(V' 9i V'~) T . . . ]T Inpi (OIOi ) 3) (11) The output from the operator in square brackets is an infinite number of derivatives arranged as a column vector. Wi is also a column vector. The expressions for the two true models can be incorporated into an optimal minimum Bayes error decision A rule as follows , where 0 priors, AT AT [0 1 , 02 ]T , W = [w i, WJjT, and b encodes the class +b wi[l, V'~1' vec(V' 91V'~1) T ... ]T Inp1 (OIOd- a w J [l , V'~,' vec(V' 92 V'~) T ... ]T Inp2(OI02) + b a T T )T w T[1, V' 9' vec ( V' 9 V' 9 . . . ]T I n P1(OIOd + b P2(OI02) w Tiplr(o) + b a Inp1(OIOd -lnp2(OI02) A a (12) iplr(o) is a score in the likelihood-ratio score-space formed by an infinite number of derivatives with respect to the parameter estimates O. Therefore, the optimal decision rule can be recovered by constructing a well-trained linear classifier in iplr(o) . In this case, the standard SVM margin can be interpreted as the log posterior margin. This justifies the use of the likelihood-ratio score-space and encourages the use of higher-order derivatives. However, most HMMs used in speech recognition are 1st-order Markov processes but speech is a high-order or infinite-order Markov process. Therefore, a linear decision boundary in the likelihood-ratio score-space defined on 1st-order Markov model estimates is unlikely to be sufficient for recovering the optimal decision rule due to model incorrectness. However, powerful non-linear classifiers may be trained in such a likelihood-ratio score-space to try to compensate for model incorrectness and approximate the optimal decision rule. SVMs with nonlinear kernels such as polynomials or Gaussian Radial Basis Functions (GRBFs) may be used. Although gains are expected from incorporating higher-order derivatives into the score-space, the size of the score-space dramatically increases. Therefore, practical systems may truncate the likelihood-ratio score-space after the 1st-order derivatives, and hence use linear approximations to the Taylor series expansions 3 . However, an example of a 2nd-order derivative is V' J-L jk (V'~;k Inp(OIO)) , T V' J-L;k (V'~;k Inp(OIO)) ~ - L 'Yjk(t)"2';;k1 (13) t= l For simplicity the component posterior 'Yj k (t) is assumed independent of J-L j k. Once the score-space has been defined, an SVM classifier can be built in the score-space. If standard linear, polynomial or GRBF kernels are used in the score-space, then the space is assumed to have a Euclidean metric tensor. Therefore, the score-space should first be whitened (i.e. decorrelated and scaled) before the standard kernels are applied. Failure to perform such score-space normalisation for a linear kernel in score-space results in a kernel similar to the Plain kernel [5]. This is expected to perform poorly when different dimensions of score-space have different dynamic ranges [2]. Simple scaling has been found to be a reasonable approximation to full whitening and avoids inverting large matrices in [2] (though for classification of single observations rather than sequences, on a different database). The Fisher kernel in [4] uses the Fisher Information matrix to normalise the score-space. This is only an acceptable normalisation for a likelihood score-space under conditions that give a zero expectation in score-space. The appropriate SVM kernel to use between two utterances O i and OJ in the normalised score-space is therefore the Normalised kernel, kN(Oi, OJ) (where ~sc is the covariance matrix in score-space), (14) 3 Experimental Results The ISOLET speaker-independent isolated letter database [8] was used for evaluation. The data was coded at a 10 msec frame rate with a 25.6 msec windowsize. The data was parameterised into 39-dimensional feature vectors including 12 MFCCs and a log energy term with corresponding delta and acceleration parameters. 240 utterances per letter from isolet{ 1,2,3,4} were used for training and 60 utterances per letter from isolet5 for testing. There was no overlap between the training and test speakers. Two sets of letters were tested, the highly confusible E-set, {B C D E G P T V Z}, and the full 26 letters. The baseline HMM system was well-trained to maximise likelihood. Each letter was modelled by a 10-emitting state left-to-right continuous density HMM with no skips, and silence by a single emitting-state HMM with no skips. Each state output distribution had the same number of Gaussian components with diagonal covariance matrices. The models were tested using a Viterbi recogniser constrained to a silence-letter-silence network. 31t is useful to note that a linear decision boundary, with zero bias, constructed in a single-dimensional likelihood-ratio score-space formed by the zeroth-order derivative operator would, under equal class priors, give the standard minimum Bayes error classifier. The baseline HMMs were used as generative models for SVM kernels. A modified version of SV Mlight Version 3.02 [9] was used to train 1vI SVM classifiers on each possible class pairing. The sequence length normalisation in Equation 10, and simple scaling for score-space normalisation, were used during training and testing. Linear kernels were used in the normalised score-space, since they gave better performance than GRBFs of variable width and polynomial kernels of degree 2 (including homogeneous, inhomogeneous, and inhomogeneous with zero-mean score-space). The linear kernel did not require parameter-tuning and, in initial experiments, was found to be fairly insensitive to variations in the SVM trade-off parameter C. C was fixed at 100, and biased hyperplanes were permitted. A variety of score-subspaces were examined. The abbreviations rn, v, wand t refer to the score-subspaces \7 J-L jk Inpi( OIOi), \7 veC (I;jk) Inpi(OIOi), \7Wjk Inpi(OIOi) and \7 ajj Inpi(OIOi) respectively. 1 refers to the log likelihood Inpi(OIOi) and r to the log likelihood-ratio In[p2(OI02) /Pl( OIOd]. The binary SVM classification results (and, as a baseline, the binary HMM results) were combined to obtain a single classification for each utterance. This was done using a simple majority voting scheme among the full set of 1v1 binary classifiers (for tied letters, the relevant 1v1 classifiers were inspected and then, if necessary, random selection performed [2]). Table 1: Error-rates for HMM baselines and SVM score-spaces (E-set) Num compo per class per state 1 2 4 6 HMM min. Bayes majority error voting 11.3 11.3 8.7 8.7 6.7 6.7 7.2 7.2 SVM lik-ratio (stat. sign.) 6.9 ~99.8~! 5.0 (98.9%) 5.7 (13.6%) 6.1 (59.5%) score-space lik lik (I-class) (2-class) 7.6 6.1 6.3 9.3 23.2 8.0 7.8 30.6 Table 1 compares the baseline HMM and SVM classifiers as the complexity of the generative models was varied. Statistical significance confidence levels are given in brackets comparing the baseline HMM and SVM classifiers with the same generative models, where 95% was taken as a significant result (confidence levels were defined by (100 - P), where P was given by McNemar's Test and was the percentage probability that the two classifiers had the same error rates and differences were simply due to random error; for this, a decision by random selection for tied letters was assigned to an 'undecided ' class [7]). The baseline HMMs were comparable to reported results on the E-set for different databases [10]. The majority voting scheme gave the same performance as the minimum Bayes error scheme, indicating that majority voting was an acceptable multi-class scheme for the E-set experiments. For the SVMs, each likelihood-ratio score-space was defined using its competing class-conditional generative models and projected into a rnr score-space. Each likelihood (I-class) score-space was defined using only the generative model for the first of its two classes, and projected into a rnl score-space. Each likelihood (2-class) score-space was defined using a generative model for both of its classes, and projected into a rnl score-space (the original Fisher score, which is a projection into its rn score-subspace, was also tested but was found to yield slightly higher error rates). SVMs built using the likelihood-ratio score-space achieved lower error rates than HMM systems, as low as 5.0%. The likelihood (I-class) score-space performed slightly worse than the likelihood-ratio score-space because it contained about half the information and did not contain the log likelihood-ratio. In both cases, the optimum number of components in the generative models was 2 per state, possibly reflecting the gender division within each class. The likelihood (2-class) score-space performed poorly possibly because of wrap-around. However, there was an excep- tion for generative models with 1 component per class per state (in total the models had 2 components per state since they modelled both classes). The 2 components per state did not generally reflect the gender division in the 2-class data, as first supposed, but the class division. A possible explanation is that each Gaussian component modelled a class with bi-modal distribution caused by gender differences. Most of the data modelled did not sit at the peaks of the two Gaussians and was not mapped to the ambiguous zero in score-subspace. Hence there was still sufficient class discrimination in score-space [7]. This task was too small to fully assess possible decorrelation in error structure between HMM and SVM classifiers [6] . Without scaling for score-space normalisation, the error-rate for the likelihood-ratio score-space defined on models with 2 components per state increased from 5.0% to 11.1%. Some likelihood-ratio mr score-spaces were then augmented with 2nd-order derivatives ~ J-t jk (~~jk lnp( 018)) . The resulting classifiers showed increases in error rate. The disappointing performance was probably due to the simplicity of the task, the independence assumption between component posteriors and component means, and the effect of noise with so few training scores in such large score-spaces. It is known that some dimensions of feature-space are noisy and degrade classification performance. For this reason, experiments were performed which selected subsets of the likelihood-ratio score-space and then built SVM classifiers in those score-subspaces. First, the score-subspaces were selected by parameter type. Error rates for the resulting classifiers, otherwise identical to the baseline SVMs, are detailed in Table 2. Again, the generative models were class-conditional HMMs with 2 components per state. The log likelihood-ratio was shown to be a powerful discriminating feature 4 ? Increasing the number of dimensions in score-space allowed more discriminative classifiers. There was more discrimination, or less noise, in the derivatives of the component means than the component variances. As expected in a dynamic task, the derivatives of the transitions were also useful since they contained some duration information. Table 2: Error rates for subspaces of the likelihood-ratio score-space (E-set) score-space error rate, % r v m mv mvt wmvtr 8.5 7.2 5.2 5.0 4.4 4.1 score-space dimensionality 1 1560 1560 3120 3140 3161 Next, subsets of the mr and wmvtr score-spaces were selected according to dimensions with highest Fisher-ratios [7] . The lowest error rates for the mr and wmvtr score-spaces were respectively 3.7% at 200 dimensions and 3.2% at 500 dimensions (respectively significant at 99.1% and 99.7% confidence levels relative to the best HMM system with 4 components per state). Generally, adding the most discriminative dimensions lowered error-rate until less discriminative dimensions were added. For most binary classifiers, the most discriminative dimension was the log likelihoodratio. As expected for the E-set, the most discriminative dimensions were dependent on initial HMM states. The low-order MFCCs and log energy term were the most important coefficients. Static, delta and acceleration streams were all useful. 4The error rate at 8.5% differed from that for the HMM baseline at 8.7% because of the non-zero bias for the SVMs. The HMM and SVM classifiers were run on the full alphabet. The best HMM classifier, with 4 components per state, gave 3.4% error rate. Computational expense precluded a full optimisation of the SVM classifier. However, generative models with 2 components per state and a wmvtr score-space pruned to 500 dimensions by Fisher-ratios, gave a lower error rate of 2.1% (significant at a 99.0% confidence level relative to the HMM system). Preliminary experiments evaluating sequence length normalisation on the full alphabet and E-set are detailed in [7]. 4 Conclusions In this work, SVMs have been successfully applied to the classification of speech data. The paper has concentrated on the nature of the score-space when handling variable length speech sequences. The standard likelihood score-space of the Fisher kernel has been extended to the likelihood-ratio score-space, and normalisation schemes introduced. The new score-space avoids some of the limitations of the Fisher score-space, and incorporates the class-conditional generative models directly into the SVM classifier. The different score-spaces have been compared on a speakerindependent isolated letter task. The likelihood-ratio score-space out-performed the likelihood score-spaces and HMMs trained to maximise likelihood. Acknowledgements N. Smith would like to thank EPSRC; his CASE sponsor, the Speech Group at IBM U.K. Laboratories; and Thorsten Joachims, University of Dortmund, for BV Mlight. References [1] V. Vapnik. The Nature of Statistical Learning Theory. Springer-Verlag, 1995. [2] N. Smith, M. Gales, and M. Niranjan. Data-dependent kernels in SVM classification of speech patterns. Tech. Report CUED/F-INFENG/TR.387, Cambridge University Eng.Dept., April 2001. [3] K. Tsuda et al. A New Discriminative Kernel from Probabilistic Models. In T.G . Dietterich, S. Becker and Z. Ghahramani, editors Advances in Neural Information Processing Systems 14, MIT Press, 2002. [4] T. Jaakkola and D. Haussler. Exploiting Generative Models in Discriminative Classifiers. In M.S. Kearns, S.A. Solia, and D.A. Cohn, editors, Advances in Neural Information Processing Systems 11 . MIT Press, 1999. [5] N. Oliver, B. Scholkopf, and A. Smola. Advances in Large-Margin Classifiers, chapter Natural Regularization from Generative Models. MIT Press, 2000. [6] S. Fine, J. Navratil, and R. Gopinath. A hybrid GMM / SVM approach to speaker identification. In Proceedings, volume 1, International Conference on Acoustics, Speech, and Signal Processing, May 2001. Utah, USA . [7] N. Smith and M. Gales. Using SVMs to classify variable length speech patterns. Tech. Report CUED/ F-INFENG/ TR.412, Cambridge University Eng.Dept., June 2001. [8] M. Fanty and R . Cole. Spoken Letter Recognition. In R.P. Lippmann, J .E. Moody, and D.S . Touretzky, editors, Neural Information Processing Systems 3, pages 220-226 . Morgan Kaufmann Publishers, 1991. [9] T. Joachims. Making Large-Scale SVM Learning Practical. In B. Scholkopf, C. Burges, and A. Smola, editors, Advances in Kernel Methods - Support Vector Learning. MIT-Press, 1999. [10] P.C. Loizou and A.S. Spanias. High-Performance Alphabet Recognition. IEEE Transactions on Speech and Audio Processing, 4(6):430-445, Nov. 1996.
2044 |@word version:2 polynomial:3 nd:3 ajj:4 eng:4 covariance:3 tr:2 initial:2 series:2 score:112 selecting:1 outperforms:1 o2:1 recovered:1 comparing:1 must:1 readily:1 jkl:1 distant:1 speakerindependent:1 discrimination:2 generative:30 selected:4 half:1 smith:4 compo:1 num:1 location:2 hyperplanes:1 simpler:1 constructed:1 rnl:2 pairing:1 scholkopf:2 expected:5 p1:4 multi:2 increasing:1 lowest:1 what:1 tying:1 interpreted:1 spoken:2 differing:1 every:1 voting:4 classifier:25 scaled:2 uk:2 before:1 maximise:7 engineering:2 zeroth:3 examined:1 hmms:11 ease:1 oiod:5 range:1 bi:1 practical:2 yj:1 testing:2 oio:5 cb2:2 projection:1 word:1 radial:1 inp:3 refers:1 confidence:4 protein:1 onto:1 selection:2 operator:9 applying:1 map:2 duration:3 simplicity:2 examines:2 rule:4 isolet:2 haussler:1 his:1 handle:3 variation:2 isolet5:1 inspected:1 homogeneous:1 us:3 recognition:10 jk:10 database:3 epsrc:1 loizou:1 region:1 spliced:1 counter:1 trade:1 highest:1 principled:1 complexity:1 cam:2 dynamic:3 trained:9 division:5 basis:1 chapter:1 derivation:1 train:1 undecided:1 forced:1 alphabet:3 sc:1 widely:1 otherwise:1 ability:1 noisy:1 sequence:16 interaction:1 fanty:1 relevant:2 poorly:2 achieve:1 supposed:1 wjk:3 exploiting:1 optimum:1 cued:2 ac:2 stat:1 ij:2 p2:5 recovering:1 skip:4 concentrate:1 inhomogeneous:2 occupies:1 require:1 suitability:1 investigation:1 preliminary:1 extension:1 pl:1 around:3 considered:1 mapping:7 bj:1 viterbi:1 tjk:2 major:1 cole:1 successfully:1 weighted:1 mit:4 gaussian:12 super:2 modified:2 rather:2 incorrectness:2 avoid:1 jaakkola:1 june:1 joachim:2 likelihood:41 tech:2 baseline:9 dependent:2 typically:1 unlikely:1 initially:1 hidden:1 issue:4 classification:10 among:1 lfor:1 proposes:1 art:1 constrained:1 fairly:1 equal:1 once:2 fuller:1 having:1 identical:1 represents:1 report:2 develops:1 jdt:1 few:1 individual:1 replaced:1 normalisation:12 highly:1 evaluation:1 mixture:3 bracket:2 tj:2 oliver:1 necessary:1 respective:1 taylor:2 euclidean:1 tsuda:1 isolated:3 increased:1 classify:2 earlier:1 column:3 entry:1 subset:2 successful:1 too:1 reported:1 kn:1 varies:1 sv:1 combined:1 st:8 density:2 peak:3 sensitivity:1 discriminating:1 international:1 probabilistic:1 off:1 together:1 moody:1 again:1 ambiguity:1 reflect:1 inaj:1 gale:3 possibly:2 worse:1 derivative:23 coefficient:1 caused:1 recogniser:1 vi:1 mv:1 stream:1 later:1 try:1 performed:5 tion:1 bayes:4 ass:1 oi:9 formed:4 square:1 variance:2 kaufmann:1 spaced:1 yield:1 climbing:1 modelled:5 identification:1 dortmund:1 touretzky:1 decorrelated:1 definition:3 failure:1 energy:2 associated:1 static:2 gain:1 proved:1 dimensionality:1 reflecting:1 appears:1 higher:4 permitted:1 modal:1 april:1 arranged:1 evaluated:1 though:1 done:1 parameterised:1 smola:2 until:1 grbf:1 cohn:1 nonlinear:1 accent:1 utah:1 effect:2 dietterich:1 contain:1 true:3 usa:1 hence:2 assigned:1 regularization:1 laboratory:1 during:1 self:2 encourages:1 width:1 ambiguous:1 speaker:5 hill:1 confusion:1 functional:1 ji:1 insensitive:1 volume:1 discussed:2 expressing:1 refer:1 significant:3 cambridge:6 vec:5 tuning:1 mfccs:2 had:3 lowered:1 whitening:1 mlight:2 posterior:11 showed:1 discard:1 termed:1 disappointing:1 verlag:1 binary:5 mcnemar:1 lnp:1 morgan:1 minimum:3 greater:1 mr:3 employed:1 signal:1 lik:3 full:6 compensate:1 jft:1 divided:1 niranjan:1 coded:1 sponsor:1 infeng:2 whitened:1 optimisation:1 metric:1 expectation:1 kernel:23 represent:1 achieved:1 addition:1 fine:1 addressed:1 publisher:1 ot:8 biased:1 probably:1 member:2 incorporates:1 variety:3 affect:1 independence:1 gave:4 competing:2 expression:1 becker:1 rnr:1 suffer:1 speech:17 speaking:1 dramatically:1 useful:8 generally:2 detailed:3 redefines:1 concentrated:1 svms:11 solia:1 percentage:1 sign:1 delta:2 popularity:1 per:14 discrete:1 group:1 gmm:1 v1:2 sum:3 wand:1 run:1 letter:12 powerful:3 reasonable:1 decision:8 acceptable:2 scaling:3 comparable:1 bv:1 constraint:1 encodes:1 nathan:1 argument:4 min:1 pruned:1 tv:1 according:1 truncate:1 slightly:2 em:1 unity:2 wi:2 making:1 classconditional:1 thorsten:1 taken:1 equation:3 adopted:1 available:1 gaussians:3 appropriate:2 mvt:1 alternative:2 jd:1 original:3 denotes:1 assumes:1 include:2 k1:1 ghahramani:1 tensor:1 added:1 occurs:3 wrapping:1 diagonal:1 kth:1 wrap:3 subspace:11 thank:1 mapped:2 hmm:21 normalise:1 majority:4 degrade:1 reason:2 assuming:2 length:12 retained:1 ratio:28 susceptible:1 yjk:2 expense:1 perform:2 observation:13 markov:4 discarded:1 defining:1 extended:1 incorporated:3 frame:1 rn:2 varied:1 wraparound:1 introduced:2 inverting:1 required:1 acoustic:5 address:1 precluded:1 pattern:2 built:4 oj:2 including:2 explanation:1 power:1 overlap:1 decorrelation:1 natural:2 hybrid:1 representing:1 scheme:9 occupancy:1 utterance:9 prior:2 acknowledgement:1 regularisation:1 occupation:1 relative:4 fully:2 discriminatively:2 limitation:3 gopinath:1 degree:1 sufficient:2 editor:4 pi:1 ibm:1 course:1 silence:3 bias:2 normalised:6 burges:1 fall:1 boundary:2 calculated:1 dimension:15 transition:5 avoids:3 plain:1 evaluating:1 adopts:1 forward:1 projected:3 far:1 emitting:3 transaction:1 approximate:1 lippmann:1 nov:1 preferred:1 dealing:1 assumed:2 discriminative:8 continuous:2 table:4 nature:3 expansion:2 complex:1 necessarily:1 constructing:1 did:4 pk:1 significance:1 noise:2 allowed:1 augmented:1 representative:1 differed:1 msec:2 lie:1 tied:2 pz:2 svm:20 sit:1 incorporating:1 vapnik:1 adding:1 justifies:1 margin:3 simply:2 contained:2 springer:1 inpi:10 gender:3 conditional:6 abbreviation:1 viewed:1 acceleration:2 fisher:14 generalisation:1 infinite:3 averaging:1 kearns:1 called:2 total:3 experimental:1 indicating:1 mark:1 support:2 dept:4 audio:1 tested:3 handling:2
1,146
2,045
Orientation-Selective aVLSI Spiking Neurons Shih-Chii Liu, J?org Kramer, Giacomo Indiveri, Tobias Delbruck, ? and Rodney Douglas Institute of Neuroinformatics University of Zurich and ETH Zurich Winterthurerstrasse 190 CH-8057 Zurich, Switzerland Abstract We describe a programmable multi-chip VLSI neuronal system that can be used for exploring spike-based information processing models. The system consists of a silicon retina, a PIC microcontroller, and a transceiver chip whose integrate-and-fire neurons are connected in a soft winner-take-all architecture. The circuit on this multi-neuron chip approximates a cortical microcircuit. The neurons can be configured for different computational properties by the virtual connections of a selected set of pixels on the silicon retina. The virtual wiring between the different chips is effected by an event-driven communication protocol that uses asynchronous digital pulses, similar to spikes in a neuronal system. We used the multi-chip spike-based system to synthesize orientation-tuned neurons using both a feedforward model and a feedback model. The performance of our analog hardware spiking model matched the experimental observations and digital simulations of continuous-valued neurons. The multi-chip VLSI system has advantages over computer neuronal models in that it is real-time, and the computational time does not scale with the size of the neuronal network. 1 Introduction The sheer number of cortical neurons and the vast connectivity within the cortex are difficult to duplicate in either hardware or software. Simulations of a network consisting of thousands of neurons with a connectivity that is representative of cortical neurons can take minutes to hours on a fast Pentium, particularly if spiking behavior is simulated. The simulation time of the network increases as the size of the network increases. We have taken initial steps in mitigating the simulation time of neuronal networks by developing a multi-chip VLSI system that can support spike-based cortical processing models. The connectivity between neurons on different chips and between neurons on the same chip are reconfigurable. The receptive fields are effected by appropriate mapping of the spikes from source neurons to target neurons. A significant advantage of these hardware simulation systems is their real-time property; the simulation time of these systems does not increase with the size of the network. In this work, we show how we synthesized orientation-tuned spiking neurons using the multi-chip system in Figure 1. The virtual connection from a selected set of neurons on Silicon retina Router (White matter) Network of neurons Orientationselective neurons Figure 1: Block diagram of a neuromorphic multi-chip system in which virtual connections from a set of neurons on a silicon retina onto another set of neurons on a transceiver chip are effected by a microcontroller. The retina communicates through the AER protocol to the PIC when it has an active pixel. The PIC communicates with the multi-neuron chip if the retina address falls into one of its stored templates. The address from the PIC is decoded by the multi-neuron transceiver. The address of the active neuron on this array can also be communicated off-chip to another receiver/transceiver. the retina to the target neurons on the multi-neuron transceiver chip is achieved with a PIC microcontroller and an asynchronous event-driven communication protocol. The circuit on this multi-neuron chip approximates a cortical microcircuit (Douglas and Martin, 1991). We explored different models that have been proposed for the generation of orientation tuning in neurons of the V1 cortical area. There have been earlier attempts to use multichip systems for creating orientation-selective neurons (Boahen et al., 1997; Whatley et al., 1997). In the present work, the receptive fields are created in a manner similar to that described in (Whatley et al., 1997). However we extend their work and quantify the tuning curves of different models. Visual cortical neurons receive inputs from the lateral geniculate nucleus (LGN) neurons which are not orientation-selective. Models for the emergence of orientation-selectivity in cortical neurons can be divided into two groups; feedforward models and feedback models. In a feedforward model, the orientation selectivity of a cortical neuron is conferred by the spatial alignment of the LGN neurons that are presynaptic to the cortical neuron (Hubel and Wiesel, 1962). In a feedback model, a weak orientation bias provided by the LGN input is sharpened by the intracortical excitatory and/or inhibitory feedback (Somers et al., 1995; Ben-Yishai et al., 1995; Douglas et al., 1995). In this work, we quantify the tuning curves of neurons created using a feedforward model and a feedback model with global inhibition. 2 System Architecture The multi-chip system (Figure 1) in this work consists of a 16 16 silicon ON/OFF retina, a PIC microcontroller, and a transceiver chip with a ring of 16 integrate-and-fire neurons and a global inhibitory neuron. All three modules communicate using the address event representation (AER) protocol (Lazzaro et al., 1993; Boahen, 1996). The communication channel signals consist of the address bits, the REQ signal, and the ACK signal. The PIC and the multi-neuron chip are both transceivers: They can both receive events and send events (Liu et al., 2001). The retina with an on-chip arbiter can only send events. Each pixel is composed of an adaptive photoreceptor that has a rectifying temporal differentiator (Kramer, 2001) in its feedback loop as shown in Figure 2. Positive temporal irradiance transients (dark-to-bright or ON transitions) and negative irradiance transients (bright-to-dark or OFF transitions) appear at two different outputs of the pixel. The outputs are then coded in the form of asynchronous binary pulses by two neurons within the pixel. These asynchronous pulses Arbiter ON REQ ON ACK neuron ON M1 OFF REQ OFF ACK neuron OFF bias M3 temporal differentiator M2    Figure 2: Pixel of the transient imager. The circuit contains a photodiode with a transistor in a a source-follower configuration with a high-gain inverting amplifier ( , ) in a negative feedback loop. A rectifying temporal differentiator in the feedback loop extracts transient ON and OFF signals. These signals go to individual neurons that generate the REQ signals to the arbiter. In this schematic, we only show the REQ and ACK signals to the X-arbiter. The duration of the ACK signal from the X-arbiter is extended within the pixel by a global refractory bias. This duration sets the refractory period of the neuron. are the request signals to the AER communication interface. A global parameter sets the minimum time (or refractory period) between subsequent pulses from the same output. Hence, the pixel can respond either with one pulse or multiple pulses to a transient. The pixels are arranged on a rectangular grid. The position of a pixel is encoded with a 4-bit column address (X address) and a 4-bit row address (Y address) as shown in Figure 3. An active neuron makes a request to the on-chip arbiter. If the neuron is selected by the arbiter, then the X and Y addresses which code the location of this neuron are placed on the output address bus of the chip. The retina then handshakes with the PIC microcontroller. The multi-neuron chip has an on-chip address decoder for the incoming events and an onchip arbiter to send events. The X address to the chip codes the identity of the neuron and the Y address codes the input synapse used to stimulate the neuron. Each neuron can be stimulated externally through an excitatory synapse or an inhibitory synapse. The excitatory neurons of this array are mutually connected via hard-wired excitatory synapses. These excitatory neurons also excite a global inhibitory neuron which in turn inhibits all the excitatory neurons. The membrane potentials of the neurons can be monitored by an on-chip scanner and the output spikes of the neurons can be monitored by the chip?s AER output. The address on the output bus codes the active neuron. In this work, the excitatory neurons on the multi-neuron chip model the orientation tuning properties of simple cells in the visual cortex and the global inhibitory neuron models an inhibitory interneuron in the visual cortex. The receptive fields of the neurons are created by configuring the connections from a subset of the source pixels on the retina onto the appropriate target neurons on the multi-neuron transceiver chip through a PIC 16C74 microcontroller. The subsets of retina pixels are determined by user-supplied templates. The microcontroller filters each retinal event to 90 deg template Inhibitory synapse Excitatory synapse 15 Y 0 deg template 0 15 0 X Figure 3: Spikes from a selected set of neurons within the two rectangular regions on the retina are mapped by the PIC onto the corresponding orientation-selective neurons on the transceiver chip. The light-shaded triangles mark the somas of the excitatory neurons and the dark-shaded triangle marks the soma of the global inhibitory neuron. Only two neurons, which are mapped for orthogonal orientations, were used in this experiment. decide if it lies in one or more of the receptive fields (RFs) of the neurons on the receiver. If it does, an event is transmitted to the appropriate receiver neuron. The typical transmission time from a spike from the sender to the receiver is about 15 s. This cycle time can be reduced by using a faster processor in place of the PIC. The retina and transceiver chips can handle handshaking cycle times on the order of 100 ns. 3 Neuron Circuit The circuit of a neuron and an excitatory synapse on the transceiver chip is shown in Figure 4. The synapse circuit (M1?M4) in the left box of the figure was originally described in (Boahen, 1996). The presynaptic spike drives the transistor M4, which acts like a switch. The bias voltages  and  set the the strength and the dynamics of the synapse. The circuit in the right box of Figure 4 implements a linear threshold integrate and fire neuron with an adjustable voltage threshold, spike pulse width and refractory period. The synaptic current  charges up the capacitance of the membrane  . When the membrane potential   exceeds a threshold voltage  , the output of the transconductance amplifier M5?M9 switches to a voltage close to  . The output of the two inverters (M10? M12 and M13?M15),  , also switches to   . The bias voltage,  , limits the current through the transconductance amplifier and the first inverter. The capacitors "!# and $ implement a capacitive divider that provides positive feedback to   . This feedback speeds up the circuit?s response and provides hysteresis to ensure that small fluctuations of   around % do not make & switch erratically. When & is high, (' is discharged through transistors M20 and M21 at a rate that is dependent on  ) . This bias voltage controls the spike?s pulse width. Once   is below  % , the transconductance amplifier switches to ground. The first inverter then switches to   but  does not immediately go to zero; it decreases linearly at a rate set by *+!# . In this way, transistor M21 is kept on, even after ,  has decreased below % . As long as the gate voltage of M21 is sufficiently high, the neuron is in its refractory period. Once transistor M21 is turned off, a new spike is generated in a time that is inversely proportional to the magnitude of - & . The spike output of the circuit is taken from the output of the first inverter. Cfb Vdd Ve M1 Vw Vspk Vdd M2 Cs Iinj Vpb Vdd M5 Vmem M4 M6 Vdd M10 M3 M7 Vthr M11 M13 M12 M14 Vout Cm M8 M9 Vdd Vrfr Cr M15 M16 Vqua M17 M18 M19 M20 Vpw Cca Vtau M21 Figure 4: Circuit diagram of an excitatory synapse (left box) connected to a linear threshold integrate-and-fire neuron (right box). Transistors M16?M19 implement a spike frequency adaptation mechanism (Boahen, 1996). A fixed amount of charge (set by   ) is dumped onto the capacitor  with every output spike. The resulting charge on  sets the current that is subtracted from the input current, and the neuron?s output frequency decreases accordingly. The voltages   and   are used to set the gain and dynamics of the integrator. 4 System Responses A rotating drum with a black and white strip was placed in front of the retina. The spike addresses and spike times generated by the retina and the multi-neuron chip at an image speed of 7.9 mm/s (or 89 pixels/s) of the rotating stimulus were recorded using a logic analyzer. The orientations of the stimuli are defined in Figure 3. Each pixel of the retina responded with only one spike to the transition of an edge of the stimulus because the refractory period of the pixel was set to 500 s. The spike addresses during the time of travel of the OFF edge of a 0 deg oriented stimulus through the entire array (Figure 5(a)) indicates that almost all the pixels along a row transmitted their addresses sequentially as the edge passed by. This sequential ordering can be seen because the stimulus was oriented slightly different from 0 deg. If the stimulus was perfectly at 0 deg, then there would be a random ordering of the pixel addresses within each row. The same observation can be made for the OFF-transient spikes recorded in response to a 90 deg oriented stimulus (Figure 5(b)). The receptive fields of two orientation-selective neurons were synthesized by mapping the OFF transient outputs of a selected set of pixels on the retina as shown in Figure 3. These two neurons have orthogonal preferred orientations. The local excitatory coupling between the neurons was disabled. There is no self excitation to each neuron so we explored only a feedforward model and a feedback model using global inhibition. We varied the size and aspect ratio of the receptive fields of the neurons by changing the template size used in the mapping of the retina spikes to the transceiver chip. The template size and aspect ratio determine the orientation responses of the neurons. The orientation response of these neurons also depends on the time constant of the neuron. On this multi-neuron chip, we do not have an explicit transistor that allows us to control the time constant. Instead, we generated a leak current through in Figure 4 by controlling the source voltage of ,   250 Retina address for OFF spikes Retina address for OFF spikes 250 200 150 100 50 0 200 150 100 50 0 7.88 7.9 7.92 7.94 7.96 7.98 8 8.02 ?50 0 50 Time (s) Time (ms) (a) (b) 100 150 Figure 5: The spike addresses from the retina were recorded when a 0 deg (a) and a 90 deg (b) oriented stimulus moved across the retina. The figure shows the time progression of the stimulated pixels (OFF spikes are marked with circles) as the 0 deg oriented stimulus (see Figure 3 for the orientation definition) passed over each row in (a). The address on the ordinate is defined as 16Y + X. A similar observation is true of (b) for the ordering of the OFF-transient spikes when each column on the retina was stimulated by the 90 deg oriented stimulus.   & . By increasing    , we decrease the time constant of the neuron. Because the neuron charges up to threshold through the summation of the incoming EPSPs, it can only spike if the ISIs of the incoming spikes are small enough. The synaptic weight determines the number of EPSPs needed to drive the neuron above threshold. We first investigated the feedforward model by using a template size of 5 7 (3 deg 4.2 deg) for one neuron and 7 5 (4.2 deg 3 deg) for the second neuron. The aspect ratio of this template was 1.4. (We have repeated the following experiments using smaller template sizes (3 5 and 1 3) and the experimental results were pretty much the same.) The time constant of the neuron and synaptic gain and strength were adjusted so that both neurons responded optimally to the stimulus. The connection from the global inhibitory neuron to the two excitatory neurons was disabled. Data was collected from the multi-neuron chip for different orientations of the drum (and hence of the stimulus). The stimulus was presented approximately 500?1000 times to the retina. Since the orientation-selective neurons responded with only 1?3 spikes every time the stimulus moved over the retina, we normalized the total number of spikes collected in these experiments to the number of stimulus presentations. The results are shown as a polar plot in Figure 6(a) for the two neurons that are sensitive to orthogonal orientations. Each neuron was more sensitive to a stimulus at its preferred orientation than the nonpreferred orientations. The neuron responded more to the orthogonal orientation than to the in-between orientations because there were a small number of retina spikes that arrived with a small ISI when the orthogonally-oriented stimulus moved across the template space of the retina (see Figure 3). We used an orientation-selective (OS) index to quantify the orien  % !#     &   !     tation selectivity of the neuron. This index is defined as   % !#      &   !     where R() is the response of the neuron. As an example, R(preferred) for neuron 5, which is sensitive to vertical orientations, is R(90)+R(270) and R(nonpreferred) is R(0)+R(180). We next investigated the feedback model. In the presence of global inhibition, the multi- 90 90 60 120 150 120 150 30 180 0 210 330 240 300 60 30 180 0 330 210 300 240 270 270 (a) (b) Figure 6: Orientation tuning curves of the two neurons in the (a) absence and (b) presence of global inhibition. The responses of the neurons were measured by the number of spikes collected per stimulus presentation. The radius of the polar plot is normalized to the maximum response of both neurons. The data was collected for stimulus orientations spaced at 30 deg intervals. The neuron that responded preferably to a 90 deg oriented stimulus (solid curve) also had a small response to a stimulus at 0 deg orientation (OS=0.428). The same observation is true for the other neuron (dashed curve) (OS=0.195). In the presence of global inhibition, each neuron responded less to the non-preferred orientation due to the suppression from the other neuron (cross-orientation inhibition). The output firing rates were also lower in this case (approximately half of the firing rates in the absence of inhibition). The OS indices are 0.546 (solid curve) and 0.497 (dashed curve) respectively. neuron system acts like a soft winner-take-all circuit. We tuned the coupling strengths between the excitatory neurons and the inhibitory neuron so that we obtained the optimal response to the same stimulus presentations as in the feedforward case. The new tuning curves are plotted in Figure 6(b). The non-preferred response of a neuron was suppressed by the other neuron through the recurrent inhibition (cross-orientation inhibition). 5 Conclusion We demonstrated a programmable multi-chip VLSI system that can be used for exploring spike-based processing models. This system has advantages over computer neuronal models in that it is real-time and the computational time does not scale with the size of the neuronal network. The spiking neurons can be configured for different computational properties. Interchip and intrachip connectivity between neurons can be programmed using the AER protocol. In this work, we created receptive fields for orientation-tuned spiking neurons by mapping the transient spikes from a silicon retina onto the neurons using a microcontroller. We have not mapped onto all the neurons on the transceiver chip because the PIC microcontroller we used is not fast enough to create receptive fields for more neurons without distorting the ISI distribution of the incoming retina spikes. We evaluated the responses of the orientation-tuned spiking neurons for different receptive field sizes and aspect ratios and also in the absence and presence of feedback inhibition. In a feedforward model, the aVLSI spiking neurons show orientation selectivity similar to digital simulations of continuous-valued neurons. Adding inhibition increased the selectivity of the spiking neurons between orthogonal orientations. We can extend the multi-chip VLSI system in this work to a more sophisticated system that supports multiple senders and multiple receivers. Such a system can be used, for example, to implement multi-scale cortical models. The success of the system in this work opens up the way for more elaborate spike-based emulations in the future. 6 Acknowledgements We acknowledge T. Horiuchi for the original design of the transceiver chip and David Lawrence for the software driver development in this work. This work was supported in part by the Swiss National Foundation Research SPP grant and the K?obler Foundation. References Ben-Yishai, R., Bar-Or, R. L., and Sompolinsky, H. (1995). Theory of orientation tuning in visual cortex. P. Natl. Acad. Sci. USA, 92(9):3844?3848. Boahen, K. A. (1996). A retinomorphic vision system. IEEE Micro, 16(5):30?39. Boahen, K. A., Andreou, A., Hinck, T., Kramer, J., and Whatley, A. (1997). Computation- and memory-based projective field processors. In Sejnowski, T., Koch, C., and Douglas, R., editors, Telluride NSF workshop on neuromorphic engineering, Telluride, CO. Douglas, R., Koch, C., Mahowald, M., Martin, K., and Suarez, H. (1995). Recurrent excitation in neocortical circuits. Science, 269(5226):981?985. Douglas, R. and Martin, K. (1991). A functional microcircuit for cat visual cortex. J. Physiol., 440:735?769. Hubel, D. and Wiesel, T. (1962). Receptive fields, binocular interaction and functional architecture. J. of Physio.(Lond), 160:106?154. Kramer, J. (2001). An integrated optical transient sensor. Submitted for publication. Lazzaro, J., Wawrzynek, J., Mahowald, M., Sivilotti, M., and Gillespie, D. (1993). Silicon auditory processors as computer peripherals. IEEE Transactions on Neural Networks, 4(3):523?528. Liu, S.-C., Kramer, J., Indiveri, G., Delbruck, T., Burg, T., and Douglas, R. (2001). Orientation-selective aVLSI spiking neurons. Neural Networks, 14(6/7):629?643. Special Issue on Spiking Neurons in Neuroscience and Technology. Somers, D., Nelson, S., and Sur, M. (1995). An emergent model of orientation selectivity in cat visual cortex simple cells. Journal of Neuroscience, 15(8):5448?5465. Whatley, A., Kramer, J., and Douglas, R. (1997). ON/OFF retina to silicon cortex. In Sejnowski, T., Koch, C., and Douglas, R., editors, Telluride NSF workshop on neuromorphic engineering, Telluride, CO.
2045 |@word wiesel:2 open:1 simulation:7 pulse:8 solid:2 initial:1 liu:3 contains:1 configuration:1 tuned:5 current:5 router:1 follower:1 physiol:1 subsequent:1 plot:2 half:1 selected:5 accordingly:1 provides:2 location:1 org:1 along:1 m7:1 driver:1 transceiver:13 consists:2 manner:1 behavior:1 isi:3 multi:23 integrator:1 m8:1 increasing:1 provided:1 matched:1 circuit:12 sivilotti:1 cm:1 temporal:4 winterthurerstrasse:1 every:2 act:2 charge:4 preferably:1 imager:1 control:2 configuring:1 grant:1 appear:1 positive:2 engineering:2 local:1 limit:1 tation:1 acad:1 fluctuation:1 firing:2 approximately:2 onchip:1 black:1 physio:1 shaded:2 co:2 programmed:1 projective:1 block:1 implement:4 communicated:1 swiss:1 area:1 eth:1 onto:6 close:1 demonstrated:1 send:3 go:2 duration:2 rectangular:2 immediately:1 m2:2 array:3 handle:1 target:3 controlling:1 user:1 us:1 synthesize:1 particularly:1 photodiode:1 vspk:1 m14:1 module:1 suarez:1 thousand:1 interchip:1 region:1 connected:3 cycle:2 sompolinsky:1 ordering:3 decrease:3 boahen:6 leak:1 tobias:1 dynamic:2 vdd:5 m18:1 req:5 triangle:2 chip:40 emergent:1 cat:2 horiuchi:1 fast:2 describe:1 sejnowski:2 neuroinformatics:1 whose:1 encoded:1 valued:2 emergence:1 advantage:3 whatley:4 transistor:7 interaction:1 erratically:1 adaptation:1 turned:1 loop:3 moved:3 m21:5 transmission:1 wired:1 ring:1 ben:2 coupling:2 avlsi:3 recurrent:2 measured:1 dumped:1 epsps:2 c:1 quantify:3 switzerland:1 radius:1 emulation:1 filter:1 transient:10 virtual:4 summation:1 adjusted:1 exploring:2 scanner:1 mm:1 around:1 sufficiently:1 ground:1 koch:3 lawrence:1 mapping:4 iinj:1 inverter:4 polar:2 geniculate:1 travel:1 sensitive:3 create:1 sensor:1 cr:1 voltage:9 publication:1 indiveri:2 indicates:1 pentium:1 suppression:1 dependent:1 entire:1 integrated:1 vlsi:5 selective:8 lgn:3 mitigating:1 pixel:19 issue:1 orientation:40 development:1 retinomorphic:1 spatial:1 special:1 field:11 once:2 future:1 stimulus:22 duplicate:1 micro:1 retina:31 oriented:8 composed:1 ve:1 national:1 individual:1 m4:3 consisting:1 fire:4 attempt:1 amplifier:4 alignment:1 light:1 natl:1 yishai:2 edge:3 orthogonal:5 rotating:2 circle:1 plotted:1 increased:1 column:2 soft:2 earlier:1 delbruck:2 neuromorphic:3 mahowald:2 subset:2 front:1 optimally:1 stored:1 giacomo:1 m10:2 off:16 multichip:1 connectivity:4 sharpened:1 recorded:3 creating:1 potential:2 m9:2 intracortical:1 retinal:1 matter:1 configured:2 hysteresis:1 depends:1 microcontroller:9 effected:3 rodney:1 rectifying:2 bright:2 responded:6 spaced:1 discharged:1 chii:1 weak:1 vout:1 drive:2 processor:3 submitted:1 synapsis:1 synaptic:3 strip:1 definition:1 frequency:2 monitored:2 gain:3 auditory:1 sophisticated:1 irradiance:2 originally:1 response:12 synapse:9 arranged:1 evaluated:1 microcircuit:3 box:4 binocular:1 o:4 stimulate:1 disabled:2 usa:1 divider:1 true:2 normalized:2 hence:2 white:2 wiring:1 during:1 width:2 self:1 excitation:2 m:1 m5:2 ack:5 arrived:1 neocortical:1 interface:1 image:1 functional:2 spiking:11 winner:2 refractory:6 analog:1 extend:2 approximates:2 m1:3 synthesized:2 silicon:8 significant:1 tuning:7 grid:1 analyzer:1 had:1 cortex:7 inhibition:11 driven:2 selectivity:6 binary:1 success:1 transmitted:2 minimum:1 seen:1 determine:1 period:5 dashed:2 signal:9 multiple:3 exceeds:1 faster:1 cross:2 long:1 divided:1 coded:1 schematic:1 vision:1 achieved:1 cell:2 receive:2 decreased:1 interval:1 diagram:2 source:4 capacitor:2 vw:1 presence:4 feedforward:8 enough:2 m6:1 switch:6 architecture:3 perfectly:1 drum:2 distorting:1 passed:2 lazzaro:2 differentiator:3 programmable:2 orientationselective:1 amount:1 dark:3 hardware:3 reduced:1 generate:1 supplied:1 nsf:2 inhibitory:10 neuroscience:2 per:1 group:1 shih:1 sheer:1 soma:2 threshold:6 changing:1 douglas:9 kept:1 v1:1 vast:1 communicate:1 respond:1 somers:2 place:1 almost:1 decide:1 bit:3 cca:1 aer:5 strength:3 software:2 aspect:4 speed:2 lond:1 transconductance:3 optical:1 martin:3 inhibits:1 developing:1 peripheral:1 request:2 membrane:3 across:2 slightly:1 smaller:1 suppressed:1 wawrzynek:1 taken:2 zurich:3 bus:2 mutually:1 turn:1 conferred:1 mechanism:1 needed:1 cfb:1 progression:1 appropriate:3 subtracted:1 gate:1 original:1 capacitive:1 ensure:1 burg:1 c74:1 capacitance:1 spike:36 receptive:10 mapped:3 simulated:1 lateral:1 decoder:1 sci:1 nelson:1 presynaptic:2 collected:4 code:4 sur:1 index:3 ratio:4 difficult:1 negative:2 design:1 adjustable:1 m11:1 m12:2 neuron:126 observation:4 vertical:1 acknowledge:1 extended:1 communication:4 varied:1 pic:12 ordinate:1 david:1 inverting:1 connection:5 andreou:1 hour:1 address:23 bar:1 below:2 spp:1 rf:1 memory:1 gillespie:1 event:10 m15:2 technology:1 inversely:1 orthogonally:1 created:4 extract:1 acknowledgement:1 vmem:1 generation:1 proportional:1 digital:3 foundation:2 integrate:4 nucleus:1 editor:2 row:4 excitatory:14 placed:2 supported:1 asynchronous:4 vthr:1 bias:6 institute:1 fall:1 template:10 feedback:13 curve:8 cortical:11 transition:3 made:1 adaptive:1 transaction:1 preferred:5 logic:1 deg:17 global:12 active:4 hubel:2 incoming:4 sequentially:1 receiver:5 excite:1 arbiter:8 continuous:2 pretty:1 stimulated:3 channel:1 m13:2 investigated:2 protocol:5 linearly:1 repeated:1 neuronal:7 representative:1 m20:2 elaborate:1 n:1 position:1 decoded:1 explicit:1 lie:1 communicates:2 externally:1 minute:1 reconfigurable:1 explored:2 consist:1 workshop:2 sequential:1 adding:1 magnitude:1 interneuron:1 sender:2 visual:6 ch:1 determines:1 identity:1 kramer:6 marked:1 presentation:3 telluride:4 absence:3 hard:1 determined:1 typical:1 total:1 experimental:2 m3:2 photoreceptor:1 nonpreferred:2 support:2 mark:2 handshaking:1
1,147
2,046
Multi Dimensional ICA to Separate Correlated Sources Roland Vollgraf, Klaus Obermayer Department of Electrical Engineering and Computer Science Technical University of Berlin Germany { vro, oby} @cs.tu-berlin.de Abstract We present a new method for the blind separation of sources, which do not fulfill the independence assumption. In contrast to standard methods we consider groups of neighboring samples ("patches") within the observed mixtures. First we extract independent features from the observed patches. It turns out that the average dependencies between these features in different sources is in general lower than the dependencies between the amplitudes of different sources. We show that it might be the case that most of the dependencies is carried by only a small number of features. Is this case - provided these features can be identified by some heuristic - we project all patches into the subspace which is orthogonal to the subspace spanned by the "correlated" features. Standard ICA is then performed on the elements of the transformed patches (for which the independence assumption holds) and robustly yields a good estimate of the mixing matrix. 1 Introduction ICA as a method for blind source separation has been proven very useful in a wide range of statistical data analysis. A strong criterion, that allows to detect and separate linearly mixed source signals from the observed mixtures, is the independence of the source signals amplitude distribution. Many contrast functions rely on this assumption, e.g. in the way, that they estimate the Kullback-Leibler distance to a (non-Gaussian) factorizing multivariate distribution [1 , 2, 3]. Others consider higher order moments of the source estimates [4, 5]. Naturally these algorithms fail when the independence assumption does not hold. In such situations it can be very useful to consider temporal/spatial statistical properties of the source signals as well. This has been done in form of suitable linear filtering [6] to achieve a sparse and independent representation of the signals. In [7] the author suggests to model the sources as a stochastic process and to do the ICA on the innovations rather than on the signals them self. In this work we extend the ICA to multidimensional channels of neighboring realizations. The used data model is explained in detail in the following section. In section 3 it will be shown, that there are optimal features, that carry lower dependencies between the sources and can be used for source separation. A heuristic is introduced, that allows to discard those features, that carry most of the dependencies. This leads to the Two-Step algorithm described in section 4. Our method requires (i) sources which exhibit correlations between neighboring pixels (e.g. continuous sources like images or sound signals), and (ii) sources from which sparse and almost independent features can be extracted. In section 5 we show separation results and benchmarks for linearly mixed passport photographs. The method is fast and provides good separation results even for sources, whose correlation coefficient is as large as 0.9. 2 Sources and observations Let us consider a set of N source signals Si(r), i = 1, ... , N of length L, where r is a discrete sample index. The sample index could be of arbitrary dimension, but we assume that it belongs to some metric space so that neighborhood relations can be defined. The sample index might be a scalar for sources which are time series and a two-dimensional vector for sources which are images 1 . The sources are linearly combined by an unknown mixing matrix A of full rank to produce a set of N observations Xi(r), N Xi(r) = l: AijSj(r) , (1) j =l and we assume that the mixing process is stationary, i.e. that the mixing matrix A is independent of r. In the following we refer to the vectors S(r) = (Sl (r), ... ,SN(r))T and X(r) = (X 1 (r), ... , XN(r))T as a source and an observation stack. The goal is to find an appropriate demixing matrix W which - when applied to the observations X(r) - recovers good estimates S(r), S(r) = WX(r) ~ S(r) (2) of the original source signals (up to a permutation and scaling of the sources). Since the mixing matrix A is not known its inverse W has to be detected blindly, i.e. only properties of the sources which are detectable in the mixtures can be exploited. For a large class of ICA algorithms one assumes that the sources are non-Gaussian and independent, i.e. that the random vector S which is sampled by L realizations S: {S(rd, 1= I, ... ,L} (3) has a factorizing and non-Gaussian joint probability distribution 2 . In situations, however, where the independence assumption does not hold, it can be helpful to take into account spatial dependencies, which can be very prominent for natural signals, and have been subject for a number of blind source separation algorithms [8, 9, 6]. Let us now consider patches si(r), s(r) = (4) 1 In the following we will mostly consider images, hence we will refer to the abovementioned neighborhood relations as spatial relations. 2In the following, symbols without sample index will refer to the random variable rather than to the particular realization. of M ? L neighboring source samples. si(r) could be a sequence of M adjacent samples of an audio signal or a rectangular patch of M pixels in an image. Instead of L realizations of a random N-vector S (cf. eq. (3)) we now obtain a little less than L realizations of a random N x M matrix s, s: {s(r)}. (5) Because of the stationarity of the mixing process we obtain x = As s and (6) = Wx, where x is an N x M matrix of neighboring observations and where the matrices A and W operate on every column vector of sand x. 3 Optimal spatial features Let us now consider a set of sources which are not statistically independent , i.e. for which N p(S) = p(Slk"'" SNk) :j:. IIp(sik) for all k = 1 ... M. (7) i=1 Our goal is to find in a first step a linear transformation 0 E IRMxM which when applied to every patch - yields transformed sources u = sOT for which the independence assumption, p(Ulk, ... ,UNk) = rr~1p(Uik) does hold for all k = 1 .. . M, at least approximately. When 0 is applied to the observations x , v = xOT , we obtain a modified source separation problem (8) where the demixing matrix W can be estimated from the transformed observations v in a second step using standard ICA. Eq. (7) is tantamount to positive transinformation of the source amplitudes. (9) where DKL is the Kullback-Leibler distance. As all elements of the patches are equally distributed, this quantity is the same for all k. Clearly, the dependencies, that are carried by single elements of the patches, are also present between whole patches, i.e. J(S1 , S2,"', SN) > O. However, since neighboring samples are correlated, it holds M J(S1 ,S2, "' ,SN ) < LJ(Slk ,S2k"",SNk ) . k=1 (10) Only if the sources where spatially white and s would consist of independent column vectors, this would hold with equality. When 0 is applied to the source patches, the trans-information between patches is not changed, provided 0 is a non-singular transformation. Neither information is introduced nor discarded by this transformation and it holds (11) For the optimal 0 now the column vectors of u = sOT shall be independent. From (10) and (11) it follows that M I(u1 ,u2, " ',uN) M = 2::I(ulk,u2k"",uNk) < 2::I(slk ,s2k"",sNk) k=1 (12) k=1 The column vectors of u are in general not equally distributed anymore, however the average trans-information has decreased to the level of information carried between the patches. In the experiments we shall see that this can be sufficiently small to reliably estimate the de-mixing matrix W. So it remains to estimate a matrix 0 that provides a matrix u with independent columns. We approach this by estimating 0 so that it provides row vectors of u that have independent elements, i.e. P(Ui) = IT;!1 P(Uik) for all i. With that and under the assumption that all sources may come from the same distribution and that there are no "cross dependencies" in u (i.e. p( Uik) is independent from p( Ujl) for k :j:. l), the independence is guaranteed also for whole column vectors of u. Thus, standard leA can be applied to patches of sources which yields 0 as the de-mixing matrix. For real world applications however , 0 has to be estimated from the observations xO T = v. It holds the relation v = Au, i.e. A only interchanges rows. So column vectors of u are independent to each other if, and only if columns of v are independent 3 . Thus, 0 can be computed from x as well. According to Eq. (12) the trans-information of the elements of columns of u has decreased in average, but not necessarily uniformly. One can expect some columns to have more independent elements than others. Thus, it may be advantageous to detect these columns rsp. the corresponding rows of 0 and discard them prior to the second leA step. Each source patch Si can be considered as linear combination of independent components, that are given by the columns of 0- 1 , where the elements of Ui are the coefficients. In the result of the leA, the coefficients have normalized variance. Therefore, those components, that have large Euklidian norm, occur as features with high entropy in the source patches. At the same time it is clear that, if there are features , that are responsible for the source dependencies, these features have to be present with large entropy, otherwise the source dependencies would have been low. Accordingly we propose a heuristic that discards the rows of 0 with the smallest Euklidian norm prior to the second leA step. How many rows have to be discarded and if this type of heuristic is applicable at all , depends of the statistical nature of the sources. In section 5 we show that for the test data this heuristic is well applicable and almost all dependencies are contained in one feature. 4 The Two-Step algorithm The considerations of the previous section give rise to a Two-Step algorithm. In the first step the transformation 0 has to be estimated. Standard leA [1, 2, 5] is performed on M -dimensional patches, which are chosen with equal probability from all of the observed mixtures and at random positions. The positions may overlap but don't overlap the boundaries of the signals. The resulting "demixing matrix" 0 is applied to the patches of observations, generating a matrix v(r) = x(r )OT, the columns of which are candidates for the input for the second leA. A number of M D columns that belong to rows of 0 with small norm are discarded as they very likely represent features , that carry dependencies between the sources. M D is chosen as a model parameter or it can be determined empirically, given the data at hand (for instance by detecting a major jump in the 3We assume non-Gaussian distributions for u and v. increase of the row norm of n). For the remaining columns it is not obvious which one represents the most sparse and independent feature. So any of them with equal probability now serve as input sample for the second ICA , which estimates the demixing matrix W. When the number N of sources is large, the first ICA may fail to extract the independent source features, because, according to the central limit theorem, the distribution of their coefficients in the mixtures may be close to a Gaussian distribution. In such a situation we recommend to apply the abovementioned two steps repeatedly. The source estimates Wx(r) are used as input for the first ICA to achieve a better n, which in turn allows to better estimate W. Figure 1: Results of standard and multidimensional ICA performed on a set of 8 correlated passport images. Top row: source images; Second row: linearly mixed sources; Third row: separation results using kurtosis optimization (FastICA Matlab package); Bottom row: separation results using multidimensional ICA (For explanation see text). 5 Numerical experiments We applied our method to a linear mixture of 8 passport photographs which are shown in Fig. 1, top row. The images were mixed (d. Fig. 1, second row) using a matrix whose elements were chosen randomly from a normal distribution with mean zero and variance one. The mixing matrix had a condition number of 80. The correlation coefficients of the source images were between 0.4 and 0.9 so that standard ICA methods failed to recover the sources: Fig. 1, 3rd row, shows the results of a kurtosis optimization using the FastICA Matlab package 4 . Fig. 1, bottom row, shows the result of the Two-Step multidimensional ICA described in section 4. For better comparison images were inverted manually to appear positive. In the first step n was estimated using FastICA on 105 patches, 6 x 6 pixels in size, which were taken with equal probability from random positions from all mixtures. The result of the first ICA is displayed in Fig. 2. The top row shows the row vectors of n sorted by the logarithm of their norm. The second row shows the features (the corresponding columns of n - 1 ) which are extracted by n . In the dia4http://www.cis.hut.fi/projects/ica/fastica/ V'Lt!1 gram below the stars indicate the logarithm of the row norm, log 0%1' and the squares indicate the mutual information J(Ulk,U7k) between the k-th features in sources 1 and 7 5, calculated using a histogram estimator. It is quite prominent that (i) a small norm of a column vector corresponds to a strongly correlated feature, and (ii) there is only one feature which carries most of the dependencies between the sources. Thus, the first column of v was discarded. The second ICA was applied to any of the remaining components, chosen randomly and with equal probability. A comparison between Figs. 1, top and bottom rows, shows that all sources were successfully recovered. Figure 2: Result of an ICA (kurtosis optimization) performed on patches of observations (cf. Fig. 1, 2nd row), 6 x 6 pixels in size. Top row: Row vectors of the demixing matrix O. Second row: Corresponding column vectors of 0- 1 . Vectors are sorted by increasing norm of the row vectors; dark and bright pixels indicate positive and negative values. Bottom diagram: Logarithm of the norm of row vectors (stars) and mutual information J(Ulk' U7k) (squares) between the coefficients of the corresponding features in the source images 1 and 7. In the next experiment we examined the influence of selecting columns of v prior to the second ICA. In Fig. 3 we show the reconstruction error (cf. appendix A), that could be achieved with the second ICA when only a single column of v served as input. From the previous experiment we have seen, that only the first component has considerable dependencies. As expected, only the first column yields poor reconstruction error. Fig. 4 shows the reconstruction error vs. M D when the M D smallest norm rows of 0 (rsp. columns of v) are discarded. We see, that for all values a good reconstruction is achieved (re < 0.6). Even if no row is discarded the result is only slightly worse than for one or two discarded rows. The dependencies of the first component are "averaged" by the vast majority of components, that carry no dependencies, in this case. The conspicuous large variance of the error for larger numbers M D might be due to convergence instabilities or close to Gaussian distributed columns of u. In either case it gives rise to discard as few components as possible. To evaluate the influence of the patch size M, the Two-Step algorithm was applied to 9 different mixtures of the sources shown in Fig. 1, top row, and using patch sizes between M = 2 x 2 and M = 6 x 6. Table 1 shows the mean and standard deviation of the achieved reconstruction error. The mixing matrix A was randomly chosen from a normal distribution with mean zero and variance one. FastICA was used for both steps, where 5.105 sample patches were used to extract the optimal features and 2.5.104 samples were used to estimate W. The smallest row of 0 was always discarded. The algorithm shows a quite robust performance, and even for patch sizes of 2 x 2 pixels a fairly good separation result is achieved 5Images no. 1 and 7 were chosen exemplarily as the two most strongly correlated sources. Jl ? ~. ==1 !C.,,". :..::. :':,. : !?;: =I 1 6 11 16 21 large row norm 26 31 36 0 5 10 small row norm Figure 3: Every single row of 0 used to generate input for the second leA. Only the first (smallest norm) row causes bad reconstruction error for the second leA step. patch size M J-lr e (Jre 2x2 3x3 4x4 5x5 6x6 0.4361 0.2322 0.1667 0.1408 0.1270 0.0383 0.0433 0.0263 0.0270 0.0460 Figure 4: M D rows with smallest norm discarded. All values of M D provide good reconstruction error in the second step. Note the slidely worse result for MD=O! Table 1: Separation result of the TwoStep algorithm performed on a set of 8 correlated passport images (d. Fig. 1, top row). The table shows the average reconstruction error J-lr e and its standard deviation (Jr e calculated from 9 different mixtures. (Note, for comparison, that the reconstruction error of the separation in Fig. 1, bottom row, was 0.2). 6 Summary and outlook We extended the source separation model to multidimensional channels (image patches). There are two linear transformations to be considered, one operating inside the channels (0) and one operating between the different channels (W). The two transformations are estimated in two adjacent leA steps. There are mainly two advantages, that can be taken from the first transformation: (i) By arranging independence among the columns of the transformed patches, the average transinformation between different channels is decreased. (ii) A suitable heuristic can be applied to discard those columns of the transformed patches, that carry most of the dependencies between different channels. A heuristic, that identifies the dependence carrying components by a small norm of the corresponding rows of 0 has been introduced. It shows, that for the image data only one component carries most of the dependencies. Due this fact, the described method works well also when all components are taken into account . In future work, we are going to establish a Maximum Likelihood model for both transformations. We expect a performance gain due to the mutual improvement of the estimates of W and 0 during the iterations. It remains to examine what the model has to be in case some rows of 0 are discarded. In this case the transformations don't preserve the dimensionality of the observation patches. A Reconstruction error The reconstruction error re is a measure for the success of a source separation. It compares the estimated de-mixing matrix W with the inverse of the original mixing matrix A with respect to the indeterminacies: scalings and permutations. It is always nonnegative and equals zero if, and only if P = W A is a nonsingular permutation matrix. This is the case when for every row of P exactly one element is different from zero and the rows of P are orthogonal, i.e. ppT is a diagonal matrix. The reconstruction error is the sum of measures for both aspects N re N N N N 2LlogL P 7j - Llog LPij 3 i=1 j=1 i=1 j=1 N N N N i=1 j=1 i=1 j=1 + Llog L P 7j -log detppT i=1 L log L P 7j - L log L pi j - N j=1 log det ppT . (13) Acknowledgment: This work was funded by the German Science Foundat ion (grant no. DFG SE 931/1-1 and DFG OB 102/3-1 ) and Wellcome Trust 061113/Z/00. References [1] Anthony J. Bell and Terrence J . Sejnowski, "An information-maximization approach to blind separation and blind deconvolution," Neural Computation, vol. 7, no. 6, pp. 1129-1159, 1995. [2] S. Amari, A. Cichocki, and H. H. Yang, "A new learning algorithm for blind signal separation," in Advances in Neural Information Processing Systems, D. S. Touretzky, M. C. Mozer, and M. E. Hasselmo, Eds., 1995, vol. 8. [3] J . F. Cardoso, "Infomax and maximum likelihood for blind source separation," IEEE Signal Processing Lett., 1997. [4] J ean-Franc;ois Cardoso, Sandip Bose, and Benjamin Friedlander, "On optimal source separation based on second and fourth order cumulants," in Proc. IEEE Workshop on SSAP, Co rfou, Greece, 1996. [5] A. Hyvarinen and E. Oja, "A fast fixed point algorithm for independent component analysis.," Neural Comput., vol. 9, pp. 1483- 1492,1997. [6] M. Zibulevski and B. A. Pearlmutter, "Blind source separation by sparse decomposition in a signal dictionary," Neural Computation, vol. 12, no. 3, pp. 863- 882, April 200l. [7] A. Hyvi:irinen, "Independent component analysis for t ime-dependent stochastic processes," in Proc. Int. Conf. on Artificial Neural Networks (ICANN'98), 1998, pp. 541-546. [8] 1. Molgedey and H. G. Schuster, "Separation of a mixture of independent signals using t ime delayed correlations," Phys. Rev. Lett., vol. 72, pp. 36343637, 1994. [9] H. Attias and C. E. Schreiner, "Blind source separation and deconvolution: The dynamic component analysis algorithm," Neural Comput., vol. 10, pp. 1373- 1424, 1998. [10] Anthony J. Bell and Terrence J. Sejnowski, "The 'independent components' of natural scenes are edge filters," Vision Res. , vol. 37, pp. 3327- 3338, 1997.
2046 |@word advantageous:1 norm:15 nd:1 decomposition:1 outlook:1 carry:7 moment:1 series:1 selecting:1 recovered:1 si:4 numerical:1 wx:3 v:1 stationary:1 accordingly:1 lr:2 detecting:1 provides:3 inside:1 expected:1 ica:20 nor:1 examine:1 multi:1 little:1 increasing:1 provided:2 project:2 estimating:1 what:1 transformation:9 temporal:1 every:4 multidimensional:5 exactly:1 grant:1 appear:1 positive:3 engineering:1 limit:1 approximately:1 might:3 au:1 examined:1 suggests:1 co:1 sandip:1 range:1 statistically:1 averaged:1 acknowledgment:1 responsible:1 x3:1 bell:2 close:2 influence:2 instability:1 www:1 rectangular:1 schreiner:1 estimator:1 spanned:1 arranging:1 element:9 observed:4 bottom:5 electrical:1 mozer:1 benjamin:1 ui:2 dynamic:1 carrying:1 serve:1 molgedey:1 joint:1 fast:2 sejnowski:2 detected:1 oby:1 artificial:1 klaus:1 neighborhood:2 u2k:1 whose:2 heuristic:7 quite:2 larger:1 otherwise:1 amari:1 sequence:1 rr:1 advantage:1 kurtosis:3 propose:1 reconstruction:12 lpij:1 tu:1 neighboring:6 realization:5 mixing:12 achieve:2 convergence:1 produce:1 generating:1 indeterminacy:1 eq:3 strong:1 c:1 ois:1 come:1 indicate:3 filter:1 stochastic:2 euklidian:2 sand:1 hold:8 sufficiently:1 considered:2 hut:1 normal:2 rfou:1 major:1 dictionary:1 smallest:5 foundat:1 proc:2 applicable:2 hasselmo:1 successfully:1 clearly:1 ujl:1 gaussian:6 always:2 modified:1 fulfill:1 rather:2 improvement:1 rank:1 likelihood:2 mainly:1 contrast:2 detect:2 helpful:1 dependent:1 lj:1 relation:4 transformed:5 going:1 germany:1 pixel:6 among:1 unk:2 spatial:4 fairly:1 mutual:3 equal:5 manually:1 x4:1 represents:1 future:1 others:2 recommend:1 few:1 franc:1 ppt:2 randomly:3 oja:1 preserve:1 ime:2 dfg:2 delayed:1 stationarity:1 mixture:10 edge:1 vro:1 orthogonal:2 logarithm:3 re:4 twostep:1 instance:1 column:26 transinformation:2 cumulants:1 maximization:1 deviation:2 fastica:5 dependency:18 combined:1 terrence:2 infomax:1 central:1 iip:1 worse:2 conf:1 account:2 de:4 star:2 coefficient:6 int:1 blind:9 depends:1 performed:5 recover:1 square:2 bright:1 variance:4 yield:4 nonsingular:1 served:1 phys:1 touretzky:1 ed:1 pp:7 obvious:1 naturally:1 recovers:1 sampled:1 gain:1 dimensionality:1 greece:1 amplitude:3 higher:1 x6:1 april:1 done:1 strongly:2 correlation:4 hand:1 trust:1 jre:1 normalized:1 hence:1 equality:1 spatially:1 leibler:2 white:1 adjacent:2 x5:1 during:1 self:1 criterion:1 prominent:2 pearlmutter:1 image:14 consideration:1 fi:1 empirically:1 jl:1 extend:1 belong:1 refer:3 rd:2 had:1 funded:1 operating:2 multivariate:1 belongs:1 discard:5 xot:1 success:1 exploited:1 inverted:1 seen:1 signal:15 ii:3 full:1 sound:1 technical:1 cross:1 roland:1 equally:2 dkl:1 vision:1 metric:1 blindly:1 histogram:1 represent:1 iteration:1 achieved:4 ion:1 lea:9 decreased:3 diagram:1 singular:1 source:61 ot:1 operate:1 subject:1 yang:1 independence:8 identified:1 attias:1 det:1 cause:1 repeatedly:1 matlab:2 useful:2 clear:1 se:1 cardoso:2 slk:3 dark:1 generate:1 sl:1 estimated:6 discrete:1 shall:2 vol:7 group:1 neither:1 vast:1 sum:1 inverse:2 package:2 fourth:1 almost:2 separation:21 patch:29 ob:1 appendix:1 scaling:2 guaranteed:1 nonnegative:1 occur:1 x2:1 scene:1 u1:1 aspect:1 department:1 according:2 combination:1 poor:1 jr:1 slightly:1 conspicuous:1 rev:1 s1:2 explained:1 xo:1 taken:3 wellcome:1 remains:2 turn:2 detectable:1 fail:2 german:1 apply:1 snk:3 appropriate:1 robustly:1 anymore:1 original:2 assumes:1 remaining:2 cf:3 top:7 establish:1 quantity:1 dependence:1 md:1 diagonal:1 abovementioned:2 obermayer:1 exhibit:1 subspace:2 distance:2 separate:2 berlin:2 majority:1 length:1 index:4 innovation:1 mostly:1 negative:1 rise:2 reliably:1 unknown:1 observation:11 discarded:10 benchmark:1 displayed:1 situation:3 extended:1 stack:1 arbitrary:1 bose:1 introduced:3 trans:3 below:1 explanation:1 ulk:4 suitable:2 overlap:2 natural:2 rely:1 s2k:2 identifies:1 carried:3 extract:3 cichocki:1 sn:3 text:1 prior:3 friedlander:1 tantamount:1 exemplarily:1 expect:2 permutation:3 mixed:4 filtering:1 proven:1 sik:1 pi:1 row:42 changed:1 summary:1 wide:1 sparse:4 distributed:3 boundary:1 dimension:1 xn:1 world:1 gram:1 calculated:2 lett:2 author:1 interchange:1 jump:1 hyvarinen:1 passport:4 sot:2 kullback:2 xi:2 don:2 factorizing:2 continuous:1 un:1 table:3 channel:6 nature:1 robust:1 ean:1 necessarily:1 anthony:2 icann:1 linearly:4 whole:2 s2:2 fig:12 uik:3 position:3 comput:2 candidate:1 third:1 theorem:1 bad:1 symbol:1 demixing:5 consist:1 deconvolution:2 workshop:1 ci:1 entropy:2 lt:1 photograph:2 likely:1 failed:1 rsp:2 contained:1 scalar:1 u2:1 irinen:1 corresponds:1 extracted:2 goal:2 sorted:2 considerable:1 determined:1 uniformly:1 llog:2 evaluate:1 audio:1 schuster:1 correlated:7
1,148
2,047
The Intelligent Surfer: Probabilistic Combination of Link and Content Information in PageRank Matthew Richardson Pedro Domingos Department of Computer Science and Engineering University of Washington Box 352350 Seattle, WA 98195-2350, USA {mattr, pedrod}@cs.washington.edu Abstract The PageRank algorithm, used in the Google search engine, greatly improves the results of Web search by taking into account the link structure of the Web. PageRank assigns to a page a score proportional to the number of times a random surfer would visit that page, if it surfed indefinitely from page to page, following all outlinks from a page with equal probability. We propose to improve PageRank by using a more intelligent surfer, one that is guided by a probabilistic model of the relevance of a page to a query. Efficient execution of our algorithm at query time is made possible by precomputing at crawl time (and thus once for all queries) the necessary terms. Experiments on two large subsets of the Web indicate that our algorithm significantly outperforms PageRank in the (human-rated) quality of the pages returned, while remaining efficient enough to be used in today?s large search engines. 1 Introduction Traditional information retrieval techniques can give poor results on the Web, with its vast scale and highly variable content quality. Recently, however, it was found that Web search results can be much improved by using the information contained in the link structure between pages. The two best-known algorithms which do this are HITS [1] and PageRank [2]. The latter is used in the highly successful Google search engine [3]. The heuristic underlying both of these approaches is that pages with many inlinks are more likely to be of high quality than pages with few inlinks, given that the author of a page will presumably include in it links to pages that s/he believes are of high quality. Given a query (set of words or other query terms), HITS invokes a traditional search engine to obtain a set of pages relevant to it, expands this set with its inlinks and outlinks, and then attempts to find two types of pages, hubs (pages that point to many pages of high quality) and authorities (pages of high quality). Because this computation is carried out at query time, it is not feasible for today?s search engines, which need to handle tens of millions of queries per day. In contrast, PageRank computes a single measure of quality for a page at crawl time. This meas- ure is then combined with a traditional information retrieval score at query time. Compared with HITS, this has the advantage of much greater efficiency, but the disadvantage that the PageRank score of a page ignores whether or not the page is relevant to the query at hand. Traditional information retrieval measures like TFIDF [4] rate a document highly if the query terms occur frequently in it. PageRank rates a page highly if it is at the center of a large sub-web (i.e., if many pages point to it, many other pages point to those, etc.). Intuitively, however, the best pages should be those that are at the center of a large sub-web relevant to the query. If one issues a query containing the word jaguar, then pages containing the word jaguar that are also pointed to by many other pages containing jaguar are more likely to be good choices than pages that contain jaguar but have no inlinks from pages containing it. This paper proposes a search algorithm that formalizes this intuition while, like PageRank, doing most of its computations at crawl time. The PageRank score of a page can be viewed as the rate at which a surfer would visit that page, if it surfed the Web indefinitely, blindly jumping from page to page. Our algorithm does something closer to what a human surfer would do, jumping preferentially to pages containing the query terms. A problem common to both PageRank and HITS is topic drift. Because they give the same weight to all edges, the pages with the most inlinks in the network being considered (either at crawl or query time) tend to dominate, whether or not they are the most relevant to the query. Chakrabarti et al. [5] and Bharat and Henzinger [6] propose heuristic methods for differentially weighting links. Our algorithm can be viewed as a more principled approach to the same problem. It can also be viewed as an analog for PageRank of Cohn and Hofmann?s [7] variation of HITS. Rafiei and Mendelzon' s [8] algorithm, which biases PageRank towards pages containing a specific word, is a predecessor of our work. Haveliwala [9] proposes applying an optimized version of PageRank to the subset of pages containing the query terms, and suggests that users do this on their own machines. We first describe PageRank. We then introduce our query-dependent, contentsensitive version of PageRank, and demonstrate how it can be implemented efficiently. Finally, we present and discuss experimental results. 2 PageRank : The Random Surfer Imagine a web surfer who jumps from web page to web page, choosing with uniform probability which link to follow at each step. In order to reduce the effect of deadends or endless cycles the surfer will occasionally jump to a random page with some small probability ?, or when on a page with no out-links. To reformulate this in graph terms, consider the web as a directed graph, where nodes represent web pages, and edges between nodes represent links between web pages. Let W be the set of nodes, N=|W|, Fi be the set of pages page i links to, and B i be the set pages which link to page i. For pages which have no outlinks we add a link to all pages in the graph1. In this way, rank which is lost due to pages with no outlinks is redistributed uniformly to all pages. If averaged over a sufficient number of steps, the probability the surfer is on page j at some point in time is given by the formula: P( j ) = 1 (1 ? ? ) P (i ) +? ? N i?B j Fi (1) For each page s with no outlinks, we set Fs={all N nodes}, and for all other nodes augment B i with s. (B i ? {s}) The PageRank score for node j is defined as this probability: PR(j)=P(j). Because equation (1) is recursive, it must be iteratively evaluated until P(j) converges. Typically, the initial distribution for P(j) is uniform. PageRank is equivalent to the primary eigenvector of the transition matrix Z: ?1? Z = (1 ? ? ) ? ? + ? M ,with ? N ? NxN ? 1 ? M ji = ? Fi ?0 ? if there is an edge from i to j (2) otherwise One iteration of equation (1) is equivalent to computing xt+1=Zxt, where xjt=P(j) at iteration t. After convergence, we have xT+1=xT, or xT=ZxT, which means xT is an eigenvector of Z. Furthermore, since the columns of Z are normalized, x has an eigenvalue of 1. 3 Directed Surfer Model We propose a more intelligent surfer, who probabilistically hops from page to page, depending on the content of the pages and the query terms the surfer is looking for. The resulting probability distribution over pages is: Pq ( j ) = (1 ? ? ) Pq? ( j ) + ? ? Pq (i ) Pq (i ? j ) (3) i?B j where Pq(i?j) is the probability that the surfer transitions to page j given that he is on page i and is searching for the query q. Pq?(j) specifies where the surfer chooses to jump when not following links. Pq(j) is the resulting probability distribution over pages and corresponds to the query-dependent PageRank score (QD-PageRankq(j) ? Pq(j)). As with PageRank, QD-PageRank is determined by iterative evaluation of equation 3 from some initial distribution, and is equivalent to the primary eigenvector of the transition matrix Zq, where Z q ji = (1 ? ? ) Pq? ( j ) + ? ? Pq (i ? j ) . Although i?B j Pq(i?j) and Pq?(j) are arbitrary distributions, we will focus on the case where both probability distributions are derived from Rq(j), a measure of relevance of page j to query q: Pq? ( j ) = Rq ( j ) ? Rq ( k ) k?W Pq (i ? j ) = Rq ( j ) ? Rq (k ) (4) k?Fi In other words, when choosing among multiple out-links from a page, the directed surfer tends to follow those which lead to pages whose content has been deemed relevant to the query (according to Rq). Similarly to PageRank, when a page?s outlinks all have zero relevance, or has no outlinks, we add links from that page to all other pages in the network. On such a page, the surfer thus chooses a new page to jump to according to the distribution Pq? (j). When given a multiple-term query, Q={q 1,q 2,?}, the surfer selects a q according to some probability distribution, P(q) and uses that term to guide its behavior (according to equation 3) for a large number of steps1. It then selects another term according to the distribution to determine its behavior, and so on. The resulting distribution over visited web pages is QD-PageRankQ and is given by 1 However many steps are needed to reach convergence of equation 3. QD ? PageRank Q ( j ) ? PQ ( j ) = ? P( q) Pq ( j ) (5) q?Q For standard PageRank, the PageRank vector is equivalent to the primary eigenvector of the matrix Z. The vector of single-term QD-PageRankq is again equivalent to the primary eigenvector of the matrix Zq. An interesting question that arises is whether the QD-PageRankQ vector is equivalent to the primary eigenvector of a matrix ZQ = ? P ( q) Z q (corresponding to the combination performed by equation 5). In q?Q fact, this is not the case. Instead, the primary eigenvector of ZQ corresponds to the QD-PageRank obtained by a random surfer who, at each step, selects a new query according to the distribution P(q). However, QD-PageRankQ is approximately equal to the PageRank that results from this single-step surfer, for the following reason. Let xq be the L2-normalized primary eigenvector for matrix Zq (note element j of xq is QD-PageRankq(j)), thus satisfying xi=Tixi. Since xq is the primary eigenvector for Zq, we have [10]: ?q, r ? Q : Z q x q ? Z q x r . Thus, to a first degree of approximation, Z q ? x r ? ?Z q x q . Suppose P(q)=1/|Q|. Consider xQ = ? P( q)x q (see equation r?Q q?Q 5). Then ? ?? ? 1 ? ? 1 1 ? ? ZQ x Q = ? ? Z ?? x q ?? = ? ?? Z q ? x r ?? ? ? ?Z q x q = ? x q = x Q ? q?Q Q q ?? q? Q q?Q n ? ?? ?Q ? Q q?Q ? r?Q ? Q q?Q and thus xQ is approximately an eigenvector for ZQ. Since xQ is equivalent to QDPageRankQ, and ZQ describes the behavior of the single-step surfer, QD-PageRankQ is approximately the same PageRank that would be obtained by using the single-step surfer. The approximation has the least error when the individual random surfers defined by Zq are very similar, or are very dissimilar. ( ) The choice of relevance function Rq(j) is arbitrary. In the simplest case, Rq(j)=R is independent of the query term and the document, and QD-PageRank reduces to PageRank. One simple content-dependent function could be Rq(j)=1 if the term q appears on page j, and 0 otherwise. Much more complex functions could be used, such as the well-known TFIDF information retrieval metric, a score obtained by latent semantic indexing, or any heuristic measure using text size, positioning, etc?. It is important to note that most current text ranking functions could be easily incorporated into the directed surfer model. 4 Scalability The difficulty with calculating a query dependent PageRank is that a search engine cannot perform the computation, which can take hours, at query time, when it is expected to return results in seconds (or less). We surmount this problem by precomputing the individual term rankings QD-PageRankq, and combining them at query time according to equation 5. We show that the computation and storage requirements for QD-PageRankq for hundreds of thousands of words is only approximately 100-200 times that of a single query independent PageRank. Let W={q1, q2, ?, qm} be the set of words in our lexicon. That is, we assume all search queries contain terms in W, or we are willing to use plain PageRank for those terms not in W. Let d q be the number of documents which contain the term q. Then S = ? d q is the number of unique document-term pairs. q?W 4 .1 Disk St o ra g e For each term q, we must store the results of the computation. We add the minor restriction that a search query will only return documents containing all of the terms 1. Thus, when merging QD-PageRankq?s, we need only to know the QD-PageRankq for documents that contain the term. Each QD-PageRankq is a vector of d q values. Thus, the space required to store all of the PageRanks is S, a factor of S/N times the query independent PageRank alone (recall N is the number of web pages). Further, note that the storage space is still considerably less than that required for the search engine?s reverse index, which must store information about all document-term pairs, as opposed to our need to store information about every unique document term pair. 4 .2 Time Requirement s If Rq(j)=0 for some document j, the directed surfer will never arrive at that page. In this case, we know QD-PageRankq(j)=0, and thus when calculating QD-PageRankq, we need only consider the subset of nodes for which Rq(j)>0. We add the reasonable constraint that Rq(j)=0 if term q does not appear in document j, which is common for most information retrieval relevance metrics, such as TFIDF. The computation for term q then only needs to consider dq documents. Because it is proportional to the number of documents in the graph, the computation of QD-PageRankq for all q in W will require O(S) time, a factor of S/N times the computation of the query independent PageRank alone. Furthermore, we have noticed in our experiments that the computation converges in fewer iterations on these smaller sub-graphs, empirically reducing the computational requirements to 0.75*S/N. Additional speedup may be derived from the fact that for most words, the sub-graph will completely fit in memory, unlike PageRank which (for any large corpus) must repeatedly read the graph structure from disk during computation. 4 .3 Empirica l Sca la bilit y The fraction S/N is critical to determining the scalability of QD-PageRank. If every document contained vastly different words, S/N would be proportional to the number of search terms, m. However, this is not the case. Instead, there are a very few words that are found in almost every document, and many words which are found in very few documents2; in both cases the contribution to S is small. In our database of 1.7 million pages (see section 5), we let W be the set of all unique words, and removed the 100 most common words3. This results in |W|=2.3 million words, and the ratio S/N was found to be 165. We expect that this ratio will remain relatively constant even for much larger sets of web pages. This means QDPageRank requires approximately 165 times the storage space and 124 times the computation time to allow for arbitrary queries over any of the 2.3 million words (which is still less storage space than is required by the search engine?s reverse index alone). 1 Google has this ?feature? as well. See http://www.google.com/technology/whyuse.html. This is because the distribution of words in text tends to follow an inverse power law [11]. We also verified experimentally that the same holds true for the distribution of the number of documents a word is found in. 3 It is common to remove ?stop? words such as the, is, etc., as they do not affect the search. 2 Table 1: Results on educrawl Query QD-PR PR Table 2: Results on WebBase Query QD-PR PR chinese association computer labs financial aid intramural maternity president office sororities student housing visitor visa Average alcoholism architecture bicycling rock climbing shakespeare stamp collecting vintage car Thailand tourism Zen Buddhism Average 5 10.75 9.50 8.00 16.5 12.5 5.00 13.75 14.13 19.25 12.15 6.50 13.25 12.38 10.25 6.75 11.38 7.38 10.75 12.50 10.13 11.50 8.45 8.45 8.43 11.53 9.13 13.15 16.90 8.63 10.68 11.88 2.93 6.88 5.75 5.03 10.68 8.68 9.75 10.38 7.99 Results We give results on two data sets: educrawl, and WebBase. Educrawl is a crawl of the web, restricted to .edu domains. The crawler was seeded with the first 18 results of a search for ? University? on Google (www.google.com). Links containing ? ?? or ? cgibin? were ignored, and links were only followed if they ended with ? .html? . The crawl contains 1.76 million pages over 32,000 different domains. WebBase is the first 15 million pages of the Stanford WebBase repository [12], which contains over 120 million pages. For both datasets, HTML tags were removed before processing. We calculated QD-PageRank as described above, using Rq(j) = the fraction of words equal to q in page j, and P(q)=1/|Q|. We compare our algorithm to the standard PageRank algorithm. For content ranking, we used the same Rq(j) function as for QDPageRank, but, similarly to TFIDF, weighted the contribution of each search term by the log of its inverse document frequency. As there is nothing published about merging PageRank and content rank into one list, the approach we follow is to normalize the two scores and add them. This implicitly assumes that PageRank and content rank are equally important. This resulted in poor PageRank performance, which we found was because the distribution of PageRanks is much more skewed than the distribution of content ranks; normalizing the vectors resulted in PageRank primarily determining the final ranking. To correct this problem, we scaled each vector to have the same average value in its top ten terms before adding the two vectors. This drastically improved PageRank. For educrawl, we requested a single word and two double word search queries from each of three volunteers, resulting in a total of nine queries. For each query, we randomly mixed the top 10 results from standard PageRank with the top 10 results from QD-PageRank, and gave them to four volunteers, who were asked to rate each search result as a 0 (not relevant), 1 (somewhat relevant, not very good), or 2 (good search result) based on the contents of the page it pointed to. In Table 1, we present the final rating for each method, per query. This rating was obtained by first summing the ratings for the ten pages from each method for each volunteer, and then averaging the individual ratings. A similar experiment for WebBase is given in Table 2. For WebBase, we randomly selected the queries from Bharat and Henzinger [6]. The four volunteers for the WebBase evaluation were independent from the four for the educrawl evaluation, and none knew how the pages they were asked to rate were obtained. QD-PageRank performs better than PageRank, accomplishing a relative improvement in relevance of 20% on educrawl and 34% on WebBase. The results are statistically significant (p<.03 for educrawl and p<.001 for WebBase using a two-tailed paired ttest, one sample per person per query). Averaging over queries, every volunteer found QD-PageRank to be an improvement over PageRank, though not all differences were statistically significant. One item to note is that the results on multiple word queries are not as positive as the results on single word queries. As discussed in section 3, the combination of single word QD-PageRanks to calculate the QD-PageRank for a multiple word query is only an approximation, made for practical reasons. This approximation is worse when the words are highly dependent. Further, some queries, such as ? financial aid? have a different intended meaning as a phrase than simply the two words ? financial? and ? aid? . For queries such as these, the words are highly dependent. We could partially overcome this difficulty by adding the most common phrases to the lexicon, thus treating them the same as single words. 6 Conclusions In this paper, we introduced a model that probabilistically combines page content and link structure in the form of an intelligent random surfer. The model can accommodate essentially any query relevance function in use today, and produces higherquality results than PageRank, while having time and storage requirements that are within reason for today? s large scale search engines. Ackno w ledg ment s We would like to thank Gary Wesley and Taher Haveliwala for their help with WebBase, Frank McSherry for eigen-help, and our experiment volunteers for their time. This work was partially supported by NSF CAREER and IBM Faculty awards to the second author. Ref erences [1] J. M. Kleinberg (1998). Authoritative sources in a hyperlinked environment. Proceedings of the Ninth Annual ACM-SIAM Symposium on Discrete Algorithms. [2] L. Page, S. Brin, R. Motwani, and T. Winograd (1998). The PageRank citation ranking: Bringing order to the web. Technical report, Stanford University, Stanford, CA. [3] S. Brin and L. Page (1998). The anatomy of a large-scale hypertextual Web search engine. Proceedings of the Seventh International World Wide Web Conference. [4] G. Salton and M. J. McGill (1983). Introduction to Modern Information Retrieval. McGraw-Hill, New York, NY. [5] S. Chakrabarti, B. Dom, D. Gibson, J. Kleinberg, P. Raghavan, and S. Rajagopalan (1998). Automatic resource compilation by analyzing hyperlink structure and associated text. Proceedings of the Seventh International World Wide Web Conference. [6] K. Bharat and M. R. Henzinger (1998). Improved algorithms for topic distillation in a hyperlinked environment. Proceedings of the Twenty-First Annual International ACM SIGIR Conference on Research and Development in Information Retrieval. [7] D. Cohn and T. Hofmann (2001). The missing link - a probabilistic model of document content and hypertext connectivity. In T. K. Leen, T. G. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing Systems 13. MIT Press, Cambridge, MA. [8] D. Rafiei and A. Mendelzon (2000). What is this page known for? Computing web page reputations. Proceedings of the Ninth International World Wide Web Conference. [9] T. Haveliwala (1999). Efficient computation of PageRank. Technical report, Stanford University, Stanford, CA. [10] G. H. Golub and C. F. Van Loan (1996). Matrix Computations. Johns Hopkins University Press, Baltimore, MD, third edition. [11] G. K. Zipf (1949). Human Behavior and the Principle of Least Effort. Addison-Wesley, Cambridge, MA. [12] J. Hirai, S. Raghaven, H. Garcia-Molina, A. Paepcke (1999). WebBase: a repository of web pages. Proceedings of the Ninth World Wide Web Conference.
2047 |@word repository:2 faculty:1 version:2 disk:2 willing:1 q1:1 accommodate:1 initial:2 contains:2 score:8 document:17 outperforms:1 current:1 com:2 must:4 john:1 shakespeare:1 hofmann:2 remove:1 treating:1 alone:3 fewer:1 selected:1 item:1 indefinitely:2 authority:1 node:7 lexicon:2 predecessor:1 symposium:1 chakrabarti:2 combine:1 introduce:1 webbase:11 bharat:3 ra:1 expected:1 behavior:4 frequently:1 underlying:1 what:2 eigenvector:10 q2:1 ended:1 formalizes:1 every:4 collecting:1 expands:1 qm:1 hit:5 scaled:1 appear:1 before:2 positive:1 engineering:1 tends:2 analyzing:1 ure:1 approximately:5 suggests:1 statistically:2 averaged:1 directed:5 unique:3 practical:1 lost:1 recursive:1 sca:1 gibson:1 significantly:1 word:28 cannot:1 graph1:1 storage:5 applying:1 www:2 restriction:1 equivalent:7 center:2 missing:1 sigir:1 assigns:1 dominate:1 financial:3 handle:1 searching:1 variation:1 president:1 mcgill:1 imagine:1 today:4 suppose:1 user:1 us:1 domingo:1 element:1 satisfying:1 database:1 winograd:1 hypertext:1 thousand:1 calculate:1 cycle:1 removed:2 principled:1 intuition:1 rq:14 jaguar:4 environment:2 asked:2 dom:1 efficiency:1 completely:1 easily:1 describe:1 query:49 choosing:2 whose:1 heuristic:3 larger:1 stanford:5 otherwise:2 richardson:1 final:2 housing:1 advantage:1 eigenvalue:1 hyperlinked:2 rock:1 propose:3 redistributed:1 ment:1 relevant:7 combining:1 normalize:1 differentially:1 scalability:2 seattle:1 convergence:2 double:1 requirement:4 motwani:1 produce:1 converges:2 help:2 depending:1 minor:1 implemented:1 c:1 empirica:1 indicate:1 qd:28 guided:1 anatomy:1 correct:1 human:3 raghavan:1 brin:2 require:1 tfidf:4 hold:1 considered:1 presumably:1 surfer:25 matthew:1 intramural:1 visited:1 weighted:1 mit:1 probabilistically:2 office:1 derived:2 focus:1 improvement:2 rank:4 greatly:1 contrast:1 dependent:6 typically:1 selects:3 issue:1 among:1 html:3 augment:1 bicycling:1 outlinks:7 proposes:2 tourism:1 development:1 equal:3 once:1 never:1 having:1 washington:2 hop:1 report:2 intelligent:4 few:3 primarily:1 modern:1 randomly:2 resulted:2 individual:3 intended:1 attempt:1 highly:6 surfed:2 evaluation:3 golub:1 mcsherry:1 compilation:1 endless:1 edge:3 closer:1 necessary:1 jumping:2 column:1 disadvantage:1 phrase:2 subset:3 uniform:2 hundred:1 successful:1 seventh:2 haveliwala:3 considerably:1 combined:1 chooses:2 st:1 person:1 international:4 siam:1 probabilistic:3 hopkins:1 connectivity:1 again:1 vastly:1 containing:9 opposed:1 zen:1 worse:1 return:2 account:1 student:1 ranking:5 performed:1 lab:1 doing:1 ttest:1 contribution:2 accomplishing:1 who:4 efficiently:1 climbing:1 none:1 published:1 reach:1 frequency:1 henzinger:3 salton:1 associated:1 stop:1 steps1:1 recall:1 car:1 improves:1 vintage:1 appears:1 wesley:2 day:1 follow:4 hirai:1 improved:3 leen:1 evaluated:1 box:1 though:1 erences:1 furthermore:2 until:1 hand:1 web:26 cohn:2 google:6 quality:7 usa:1 effect:1 dietterich:1 normalized:2 true:1 contain:4 seeded:1 read:1 iteratively:1 buddhism:1 semantic:1 during:1 skewed:1 ackno:1 hill:1 demonstrate:1 performs:1 meaning:1 recently:1 fi:4 common:5 ji:2 empirically:1 million:7 analog:1 he:2 association:1 discussed:1 significant:2 distillation:1 cambridge:2 zipf:1 automatic:1 similarly:2 pointed:2 pq:17 etc:3 add:5 something:1 own:1 reverse:2 occasionally:1 store:4 molina:1 greater:1 additional:1 somewhat:1 determine:1 multiple:4 reduces:1 technical:2 positioning:1 retrieval:7 equally:1 visit:2 award:1 paired:1 xjt:1 essentially:1 metric:2 volunteer:6 blindly:1 iteration:3 represent:2 baltimore:1 source:1 unlike:1 bringing:1 tend:1 enough:1 affect:1 fit:1 gave:1 architecture:1 reduce:1 whether:3 effort:1 f:1 returned:1 york:1 nine:1 repeatedly:1 ignored:1 thailand:1 ten:3 simplest:1 http:1 specifies:1 nsf:1 per:4 discrete:1 four:3 verified:1 vast:1 graph:6 fraction:2 surmount:1 inverse:2 arrive:1 almost:1 reasonable:1 followed:1 hypertextual:1 annual:2 occur:1 constraint:1 tag:1 kleinberg:2 relatively:1 speedup:1 department:1 according:7 combination:3 poor:2 describes:1 smaller:1 remain:1 visa:1 intuitively:1 restricted:1 pr:5 indexing:1 equation:8 resource:1 discus:1 needed:1 know:2 addison:1 eigen:1 assumes:1 remaining:1 include:1 top:3 calculating:2 invokes:1 chinese:1 noticed:1 question:1 primary:8 md:1 traditional:4 link:18 zxt:2 thank:1 topic:2 reason:3 index:2 reformulate:1 ratio:2 preferentially:1 frank:1 twenty:1 perform:1 datasets:1 looking:1 incorporated:1 inlinks:5 ninth:3 arbitrary:3 drift:1 rating:4 introduced:1 pair:3 required:3 optimized:1 engine:10 hour:1 hyperlink:1 rajagopalan:1 pagerank:56 memory:1 belief:1 power:1 critical:1 difficulty:2 improve:1 pedrod:1 rated:1 technology:1 carried:1 deemed:1 tresp:1 xq:6 text:4 l2:1 determining:2 relative:1 nxn:1 law:1 expect:1 mixed:1 interesting:1 proportional:3 authoritative:1 degree:1 sufficient:1 dq:1 editor:1 principle:1 ibm:1 supported:1 drastically:1 bias:1 guide:1 allow:1 wide:4 taking:1 van:1 overcome:1 plain:1 calculated:1 crawl:6 transition:3 world:4 computes:1 ignores:1 author:2 made:2 jump:4 citation:1 implicitly:1 mcgraw:1 corpus:1 summing:1 knew:1 xi:1 search:22 iterative:1 latent:1 reputation:1 zq:10 tailed:1 table:4 ca:2 career:1 requested:1 complex:1 domain:2 precomputing:2 alcoholism:1 edition:1 nothing:1 ref:1 ny:1 aid:3 sub:4 stamp:1 weighting:1 third:1 formula:1 specific:1 xt:5 hub:1 meas:1 list:1 normalizing:1 merging:2 adding:2 execution:1 garcia:1 simply:1 likely:2 contained:2 partially:2 pedro:1 corresponds:2 gary:1 acm:2 ma:2 viewed:3 towards:1 content:12 feasible:1 experimentally:1 loan:1 determined:1 uniformly:1 reducing:1 averaging:2 total:1 experimental:1 la:1 latter:1 arises:1 dissimilar:1 relevance:7 crawler:1 visitor:1
1,149
2,048
Motivated Reinforcement Learning Peter Dayan Gatsby Computational Neuroscience Unit 17 Queen Square, London, England, WClN 3AR. dayan@gatsby.ucl.ac.uk Abstract The standard reinforcement learning view of the involvement of neuromodulatory systems in instrumental conditioning includes a rather straightforward conception of motivation as prediction of sum future reward. Competition between actions is based on the motivating characteristics of their consequent states in this sense. Substantial, careful, experiments reviewed in Dickinson & Balleine, 12,13 into the neurobiology and psychology of motivation shows that this view is incomplete. In many cases, animals are faced with the choice not between many different actions at a given state, but rather whether a single response is worth executing at all. Evidence suggests that the motivational process underlying this choice has different psychological and neural properties from that underlying action choice. We describe and model these motivational systems, and consider the way they interact. 1 Introduction Reinforcement learning (RL 28) bears a tortuous relationship with historical and contemporary ideas in classical and instrumental conditioning. Although RL sheds important light in some murky areas, it has paid less attention to research concerning the motivation of stimulus-response (SR) links. RL methods are mainly concerned with preparatory Pavlovian (eg secondary) conditioning, and, in instrumental conditioning, the competition between multiple possible actions given a particular stimulus or state, based on the future rewarding or punishing consequences of those actions. These have been used to build successful and predictive models of the activity of monkey dopamine cells in conditioning. 22,24 By contrast, SR research starts from the premise that, in many circumstances, given an unconditioned stimulus (US; such as a food pellet), there is only one natural set of actions (the habit of approaching and eating the food), and the main issue is whether this set is worth executing (yes, if hungry, no if sated). This is traditionally conceived as a question of consummatory motivation. SR research goes on to study how these habits, and also the motivation associated with them, are 'attached' in an appropriately preparatory sense to conditioned stimuli (CSs) that are predictive of the USs. The difference between RL's competition between multiple actions and SR's motivation of a single action might seem trivial, particularly if an extra, nUll, action is included in the action competition in RL, so the subject can actively choose to do nothing. However, there is substantial evidence from experi- ments in which drive states (eg hunger, thirst) are manipulated, that motivation in the SR sense works in a sophisticated, intrinsically goal-sensitive, way and can exert unexpected effects on instrumental conditioning. By comparison with RL, psychological study of multiple goals within single environments is quite advanced, particularly in experiments in which one goal or set of goals is effective during learning, and another during performance. Based on these and other studies, (and earlier theoretical ideas from, amongst others, Konorski, 18,19 Dickinson, Balleine and their colleagues 13 have suggested that there are really two separate motivational systems, one associated with Pavlovian motivation, as in SR, and one associated with instrumental action choice. They further suggest, partly based on related suggestions by Berridge and his colleagues,? that only the Pavlovian system involves dopamine. Neither the Pavlovian nor the instrumental system maps cleanly onto the standard view of RL, and the suggestion about dopamine would clearly Significantly damage the RL interpretation of the involvement of this neuromodulatory system in conditioning. In this paper, we describe some of the key evidence supporting the difference between instrumental and Pavlovian motivation (see also Balkenius 3 and Spier 25 ), and expand the model of RL in the brain to incorporate SR motivation and concomitant evidence on intrinsic goal sensitivity (as well as intrinsic habits). Some of the computational properties of this new model turn out to be rather strange - but this is a direct consequence of equivalently strange observable behavior. 2 Theoretical and Experimental Background Figure 1 shows a standard view of the involvement of the dopamine system in RL. 22 ,24 Dopamine neurons in the ventral tegmental area (VTA) and substantia nigra pars compacta (SN c) report the temporal difference (TD) error 8 (t). In the simplest version of the theory, this is calculated as 8 (t) = r (t) + V 1T (x( t + 1? V 1T(x( t) ), where r (t) is the value of the reward at time t, x( t) is an internal representation of the state at time t, V 1T(x( t? is the expectation of the sum total future reward expected by the animal based on starting from that state, following policy IT, and the transition from x( t) to x( t + 1) is occasioned by the action a selected by the subject. In the actor-critic 6 version of the dopamine theory, this TD error signal is put to two uses. One is adapting parameters that underlie the actual predictions V 1T(x(t?. For this, 8(t) > 0 if the prediction from the state at time t, V 1T(x(t?, is overly pessimistic with respect to the sum of the actual reward, r (t), and the estimated future reward, V 1T (x( t + 1?, from the subsequent state. The other use for 8 (t) is criticizing the action a adopted at time t. For this, 8(t) > 0 implies that the action chosen is worth more than the average worth of x(t), and that the overall policy IT of the subject can therefore be improved by choosing it more often. In a Q-Iearning 31 version of the theory, Q1T(X, a) values are learned using an analogous quantity to 8 (t), for each pair of states x and actions a, and can directly be used to choose between the actions to improve the policy. Even absent an account for intrinsic habits, three key paradigms show the incompleteness of this view of conditioning: appetitive Pavlovian-instrumental transfer,15 intrinsic drive preference under speCific deprivation states,8 and incentive learning, as in the control of chains of instrumental behavior. 5 The SR view of conditioning places its emphasis on motivational control of a prepotent action. That is, the natural response associated with a stimulus (presumably as output by an action specification mechanism) is only elicited if A x state B ~muli x amygdala TO prediction accumbens~--+--'--=--l error OFC reward '" '" action TO predictio error 8 r state stimuli competition stRkfu~ a Figure 1: Actor-critic version of the standard RL model. A) Evaluator: A TD error signal 8 to learn V 1T (x) to match the sum of future rewards r, putatively via the basolateral nuclei of the amygdala, the orbitofrontal cortex and the nucleus accumbens. B) Instrumental controller: The TD error 8 is used to choose, and teach the choice of, appropriate actions a to maximize future reward based on the current state and stimuli, putatively via the dorsolateral prefrontal cortex and the dorsal striatum. it is motivationally appropriate, according to the current goals of the animal. The suggestion is that this is mediated by a separate motivational system. USs have direct access to this system, and CSs have learned access. A conclusion used to test this structure for the control of actions is that this motivational system could be able to energize any action being executed by the animal. Appetitive Pavlovian-instrumental transfer 15 shows exactly this. Animals executing an action for an outcome under instrumental control, will perform more quickly when a CS predictive of reward is presented, even if the CS predicts a completely different reward from the instrumental outcome. This effect is abolished by lesions of the shell of the nucleus accumbens,10 one of the main targets of DA from the VTA. The standard RL model offers no account of the speed or force of action (though one could certainly imagine various possible extensions), and has no natural way to accommodate this finding. * The second challenge to RL comes from experiments on the effects of changing speCific and general needs for animals. For instance, Berridge & Schulkin8 first gave rats sucrose and saline solutions with one of a bitter (quinine) and a sour (citric) taste. They then artificially induced a strong physiological requirement for salt, for the first time in the life of the animal. Presented with a choice between the two flavors (in plain water, ie in extinction), the rats preferred to drink the flavor associated with the salt. Furthermore, the flavor paired with the salt was awarded positive hedonic reactions, whereas before pairing (and if it had been paired with sucrose instead) it was treated as being aversive. The key feature of this experiment is that this preference is evident without the opportunity for learning. Whereas the RL system could certainly take the physiological lack of salt as helping determine part of the state x(t), this could only exert an effect on behavior through learning, contrary to the evidence. The final complexity for standard RL comes from incentive learning. One paradigm involves a sequential chain of two actions (a1 and a2) that rats had to execute in order to get a reward. 5 The subjects were made hungry, and were first trained to perform action a2 to get a particular reward (a Noyes pellet), and then to perform the chain of actions a1 - a2 to get the reward. In a final test phase, the animals were offered the chance of executing a1 and a2 in extinction, for half of them when they were still hungry; for the other half when they were sated on their normal diet. Figure 2A shows what happens. Sated animals perform a1 at the same rate as hungry animals, but perform a2 sig*Note that aversive Pavlovian instrumental transfer, in the form of the suppression of appetitive instrumental responding, is the conventional method for testing aversive Pavlovian conditioning. There is an obvious motivational explanation for this as well as the conventional view of competition between appetitive and protective actions. 100 .~ hungry 80 ieo s a ~ 40 E20 al al al al Figure 2: Incentive learning. A) Mean total actions al and a 2 for an animal trained on the chain schedule al - a2 -Noyes pellets. Hungry and sated rats perform al at the same rate, but sated animals fail to perform a 2. B) Mean total actions when sated following prior re-exposure to the Noyes pellets when hungry ('hungry-sated') or when sated (,sated-sated'). Animals re-exposed when sated are significantly less willing to perform a 2. Note the change in scale between A and B. Adapted from Balieine et a/. 5 nificantly less frequently. Figure 2B shows the basic incentive learning effect. Here, before the test, animals were given a limited number of the Noyes pellets (without the availability of the manipulanda) either when hungry or when sated. Those who experienced them hungry ('hungry-sated') show the same results as the 'sated' group of figure 2A; whereas those who experienced them sated (,sated-sated') now declined to perform action al either. This experiment makes two points about the standard RL model. First, the action nearest to the reward (a2) is affected by the deprivation state without additional learning. This is like the effect of specific deprivation states discussed above. Second is that a change in the willingness to execute al happens after re-exposure to the Noyes pellets whilst sated; this learning is believed to involve insular cortex (part of gustatory neocortex4). That re-exposure directly affects the choice of al suggests that the instrumental act is partly determined by an evaluation of its ultimate consequence, a conclusion that relates to a long-standing psychological debate about the 'cognitive' evaluation of actions. Dickinson & Balleine 13 suggest that the execution of a2 is mainly controlled by Pavlovian contingencies, and that Pavlovian motivation is instantly sensitive to goal devaluation via satiation. At this stage in the experiment, however, al is controlled by instrumental contingencies. By comparison with Pavlovian motivation, instrumental motivation is powerful (since it can depend on response-outcome expectancies), but dumb (since, without re-exposure, the animal works hard doing al when it wouldn't be interested in the food in any case). Ultimately, after extended training,14 in the birth of a new habit, al becomes controlled by Pavlovian contingencies too, and so becomes directly sensitive to devaluation. t 3 New Model These experiments suggest some major modifications to the standard RL view. Figure 3 shows a sketch of the new model, whose key principles include ? Pavlovian motivation (figure 3A) is associated with prediction error 8(t) = r(t) + VTT(X(t + 1)) - VTT(x(t) for long term expected future rewards VTT(X, a), given a policy IT. Adopting this makes the model account for the classical conditioning paradigms explained by the standard RL model. t It is not empirically clear whether actions that have become habits are completely automatic 1 or are subject to Pavlovian motivational influences. A B stimuli X ! CS prior , bias S S ~amygdala ," : CS ~ sta~e~ __>~ stimuli X sIs US 8 C prior bias habit stimuli X CS sIs shell ~~.~ US a I s~a~: vrT(X) Figure 3: Tripartite model. A) Evaluator: USs are evaluated by a hard-wired evaluation system (HE) which is intrinsically sensitive to devaluation. USs can also be evaluated via a plastic route, as in figure 1, but which nevertheless has prior biases. CSs undergo Pavlovian stimulus substitution with the USs they predict, and can also be directly evaluated through the learned route. The two sources of information for VTT(x) compete, forcing the plastic route to adjust to the hard-wired route. B) Habit system: The SR mapping suggests an appropriate action based on the state X; the vigor of its execution is controlled by dopaminergic 8, putatively acting via the shell of the accumbens. C) Instrumental controller: Action choice is based on advantages, which are learned, putatively via the core of the accumbens. Prefrontal working memory is used to unfold the consequences of chosen actions. ? ret) is determined by a devaluation-sensitive, 'hard-wired', US evaluator that provides direct value information about motivationally inappropriate USs. ? 8(t), possibly acting through the shell of the accumbens, provides Pavlovian motivation for pre-wired and new habits (figure 3B), as in Pavlovian instrumental transfer. ? V 1T (x(t) is determined by two competing sources: one as in the standard model (involving the basolateral nuclei of the amygdala and the orbitofrontal cortex (OFC),16,23 and including prior biases (sweet tasting foods are appetitive) expressed in the connections from primary taste cortex to oFC and the amygdala; the other, which is primary, dependent largely on a stimulus substitution20 relationship between CSs and USs, that is also devaluationdependent. The latter is important for ultimate Pavlovian control over actions; the former for phenomena such as secondary conditioning, which are known to be devaluation independentY Figure 4A (dashed) shows the contribution of the hard-wired evaluation route, via stimulus-substitution, on the prediction of value in classical conditioning. Here, stimulus-substitution was based on a form of Hebbian learning with a synaptic trace, so the shorter the CS-US interval, the greater the HE component. This translates into greater immediate sensitivity to devaluation, the main characteristic of the hard-wired route. The plastic route via the amygdala takes responsibility for the remainder of the prediction; and the sum prediction is always correct (solid line). ? Short-term storage of predictive stimuli in prefrontal working memory is gated9 by 8(t), so can also be devaluation dependent. ? Instrumental motivation depends on policy-based advantages (3C; Baird 2) A 1T (x,a) = Q 1T(x,a) - V1T(X) trained by the error signal 8 A (t) = 8(t) -A1T(x,a) Over the course of policy improvement, the advantage of a sub-optimal action becomes negative, and of an optimal action tends to O. The latter offers a natural model of the transition from instrumental action selection to an SR habit. Note that, in this actor-critic scheme, some aspects of advantages are not necessary, such as the normalizing updates. B A 0.5 1:: bO 0[j \ \ ;s: 0.5 79 8 C value V7T (xo) -,-7T--,(:..:: XO,-,-,.:..:. a ,-) -----, 1 .5 ,----"'-"-''-----''-''''-'---, ~4 ,------:.A advantages sum HE \ component ' ~~~~ ,, , , -0.5 A 7T(X , b) , ?0~--1-0~~20 0 50 cs-us interval successor sum 0.2 iteration 100 o 'HE , -reprn --- 1 :- , ~~~ , ??. plastic component -0.2 50 iteration 101 0~--50 ---' 1 00 iteration Figure 4: A) Role of the hard-wired route (dashed line), via stimulus-substitution, in predicting future reward (r = 1) as a function of the CS-US interval. The solid line shows that the net prediction is always correct. B) Advantages of useful (a) and worthless (b) actions at the start state Xo. C) Evolution of the value of Xo over learning. The solid line shows the mean value; the dashed line the hard-wired component, providing immediate devaluation sensitivity. 0) Construction of A7T (xo, a) via a successor representation component l l (dashed) and a conventionally learned component (dotted). The former is sensitive to re-exposure devaluation, as in figure 2B. B-D) Action a produces reward r = 1 with probability 0.9 after 3 timesteps; curves are averages over 2000 runs. Figure 4B;C show two aspects of instrumental conditioning. Two actions compete at state xo, one, a, with a small cost and a large future payoff; the other, b, with no cost and no payoff. Figure 4B shows the development of the advantages of these actions over learning. Action a starts looking worse, because it has a greater immediate cost; its advantage increases as the worth of a grows greater than the mean value of xo, and then goes to 0 (the birth of the habit) as the subject learns to choose it every time. Figure 4C shows the value component of state Xo. This comes to be responsible for the entire prediction (as A1T(XO, a) ~ 0). As in figure 4A, there is a hard-wired component to this value which would result in the immediate decrement of response evident in figure 2A. ? On-line action choice is dependent on 8A (t) as in learned klinokinesis.21 Incentive learning in chains suggests that the representation underlying the advantage of an action includes information about its future consequences, either through an explicit model,27,29 a successor representation,ll or perhaps a form of f3-model. 26 One way of arranging this would use a VTE-like 30 mechanism for proposing actions (perhaps using working memory in prefrontal cortex), in order to test their advantages. Figure 4D shows the consequence of using a learned successor representation underlying the advantage A1T(XO, a) shown in figure 4B. The dashed line shows the component of A1T(XO, a) dependent on a learned successor representation, and the prior bias about the value of the reward, and which is therefore sensitive to re-exposure (when the value accorded to the reward is decreased); the dotted line shows the remaining component of A 1T(XO, a), learned in the standard way. Re-exposure sensitivity (ie incentive learning) will exist over roughly iterations 25 - 75 . ? SR models also force consideration of the repertoire of possible actions or responses available at a given state (figure 3B;C). We assume that both corticocortical and cortico-(dorsal) striatal plasticity sculpt this collection, using 8 A (t) directly, and maybe also correlational learning rules. The details of the model are not experimentally fully determined, although its general scheme is based quite straightforwardly from the experimental evidence referred to (and many other experiments), and by consistency with the activity of dopamine cells (recordings of which have so far used only a single motivational state). 4 Discussion Experiments pose a critical challenge to our understanding of the psychological and neural implementation of reinforcement learning, 12,13 suggesting the importance of two different sorts of motivation in controlling behavior. With both empirical and theoretical bases, we have put these two aspects together through the medium of advantages. The most critical addition is a hard-wired, stimulus-substitution sensitive, route for the evaluation of stimuli and states, which competes with a plastic route through the amygdala and the oFC. This hard-wired route has the property of intrinsic sensitivity to various sorts of devaluation, and this leads to motivationally appropriate behavior. The computational basis of the new aspects of the model focus on motivational control of SR links (via VTT ), to add to motivational control of instrumental actions (via ATT). We also showed the potential decomposition of the advantages into a component based on the successor representation and therefore sensitive to re-exposure as in incentive learning, and a standard, learned, component. The model is obviously incomplete, and requires testing in richer environments. In particular, we have yet to explore how habits get created from actions as the maximal advantage goes to o. Acknowledgements I am very grateful to Christian Balkenius, Bernard Balleine, Tony Dickinson, Sham Kakade, Emmet Spier and Angela Yu for discussions. Funding was from the Gatsby Charitable Foundation. References [1] Adams, CD (1982) Variations in the sensitivity of instrumental responding to reinforcer devaluation. QJEP 34B:77-98. [2] Baird, LC (1993) Advantage Updating. Technical report WL-TR-93-1146, Wright-Patterson Air Force Base. [3] Balkenius, C (1995) Natural Intelligence in Artificial Creatures. PhD Thesis, Department of Cognitive Science, Lund University, Sweden. [4] Balleine, BW & Dickinson, A (1998) Goal-directed instrumental action: Contingency and incentive learning and their cortical substrates. Neuropharmacology 37:407-419. [5] Balleine, BW, Garner, C, Gonzalez, F & Dickinson, A (1995) Motivational control of heterogeneous instrumental chains. Journal of Experimental Psychology: Animal Behavior Processes 21:203-217. [6] Barto, AG, Sutton, RS & Anderson, CW (1983) Neuronlike elements that can solve difficult learning problems. IEEE SMC 13:834-846. [7] Berridge, KC (2000) Reward learning: Reinforcement, incentives, and expectations. In DL Medin, editor, The Psychology of Learning and Motivation 40:223-278. [8] Berridge, KC & Schulkin, J (1989) Palatability shift of a salt-associated incentive during sodium depletion. Quarterly Journal of Experimental Psychology: Comparative & Physiological Psychology 41:121-138. [9] Braver, TS, Barch, DM & Cohen, JD (1999) Cognition and control in schizophrenia: A computational model of dopamine and prefrontal function. Biological Psychiatry 46:312-328. [10] Corbit, LH, Muir, JL & Balleine, BW (200l) The role of the nucleus accumbens in instrumental conditioning: Evidence of a functional dissociation between accumbens core and shell. Journal of Neuroscience 21:3251-3260. [11] Dayan, P (1993) Improving generalisation for temporal difference learning: The successor representation. Neural Computation 5:6l3-624. [12] Dickinson, A & Balleine, B (1994) Motivational control of goal-directed action. Animal Learning & Behavior 22:1-18. [l3] Dickinson, A & Balleine, B (200l) The role of learning in motivation. In CR Gallistel, editor, Learning, Motivation and Emotion, Volume 3 of Steven's Handbook of Experimental Psychology, Third Edition. New York, NY: Wiley. [14] Dickinson, A, Balleine, B, Watt, A, Gonzalez, F & Boakes, RA (1995) Motivational control after extended instrumental training. Animal Learning & Behavior 23:197-206. [15] Estes, WK (1943). Discriminative conditioning. I. A discriminative property of conditioned anticipation. JEP 32:150-155. [16] Holland, PC & Gallagher, M (1999) Amygdala circuitry in attentional and representational processes. Trends in Cognitive Sciences 3:65-73. [17] Holland, PC & Rescorla, RA (1975) The effect of two ways of devaluing the unconditioned stimulus after first- and second-order appetitive conditioning. Journal of Experimental Psychology: Animal Behavior Processes 1:355-363. [18] Konorski, J (1948) Conditioned Reflexes and Neuron Organization. Cam- bridge, England: Cambridge University Press. [19] Konorski, J (1967) Integrative Activity of the Brain: An Interdisciplinary Approach. Chicago, 11: University of Chicago Press. [20] Mackintosh, NJ (1974) The Psychology of Animal Learning. New York, NY: Academic Press. TJ (1995) Bee foraging in uncertain environments using predictive hebbian learning. Nature [21] Montague, PR, Dayan, P, Person, C & Sejnowski 377:725-728. [22] Montague, PR, Dayan, P & Sejnowski, TJ (1996) A framework for mesencephalic dopamine systems based on predictive Hebbian learning. Journal of Neuroscience 16: 1936-1947. [23] Schoenbaum, G, Chiba, AA & Gallagher, M (1999) Neural encoding in or- [24] [25] [26] [27] bitofrontal cortex and basolateral amygdala during olfactory discrimination learning. Journal of Neuroscience 19:1876-1884. Schultz, W, Dayan, P & Montague, PR (1997) A neural substrate of prediction and reward. Science 275:1593-1599. Spier, E (1997) From Reactive Behaviour to Adaptive Behaviour. PhD Thesis, Balliol College, Oxford. Sutton, RS (1995) TD models: modeling the world at a mixture of time scales. In A Prieditis & S Russell, editors, Proceedings of the Twelfth International Conference on Machine Learning. San Francisco, CA: Morgan Kaufmann, 531-539. Sutton, RS & Barto, AG (1981) An adaptive network that constructs and uses an internal model of its world. Cognition and Brain Theory 4:217- 246. [28] Sutton, RS & Barto, AG (1998) Reinforcement Learning. Cambridge, MA: MIT Press. [29] Sutton, RS & Pinette, B (1985) The learning of world models by connectionist networks. Proceedings of the Seventh Annual Conference of the Cognitive Science Society. Irvine, CA: Lawrence Erlbaum, 54-64. [30] Tolman, EC (1938) The determiners of behavior at a choice point. Psychological Review 45:1-41. [31] Watkins, CJCH (1989) Learning from Delayed Rewards. PhD Thesis, Uni- versity of Cambridge, Cambridge, UK.
2048 |@word nificantly:1 version:4 instrumental:29 extinction:2 twelfth:1 cleanly:1 willing:1 r:5 integrative:1 decomposition:1 paid:1 tr:1 solid:3 accommodate:1 substitution:5 att:1 reaction:1 current:2 si:2 yet:1 subsequent:1 chicago:2 plasticity:1 christian:1 update:1 discrimination:1 half:2 selected:1 intelligence:1 core:2 short:1 provides:2 putatively:4 preference:2 evaluator:3 direct:3 become:1 gallistel:1 pairing:1 olfactory:1 balleine:10 ra:2 expected:2 vtt:5 roughly:1 nor:1 frequently:1 behavior:10 brain:3 v1t:1 preparatory:2 td:5 food:4 actual:2 versity:1 inappropriate:1 becomes:3 motivational:14 underlying:4 v7t:1 competes:1 medium:1 null:1 what:1 monkey:1 accumbens:8 proposing:1 whilst:1 ret:1 finding:1 ag:3 sour:1 nj:1 cjch:1 temporal:2 every:1 act:1 shed:1 iearning:1 exactly:1 uk:2 control:11 unit:1 underlie:1 positive:1 before:2 tends:1 consequence:6 striatum:1 sutton:5 encoding:1 insular:1 oxford:1 might:1 exert:2 emphasis:1 tasting:1 suggests:4 balkenius:3 limited:1 smc:1 medin:1 directed:2 responsible:1 testing:2 jep:1 habit:12 substantia:1 unfold:1 area:2 empirical:1 significantly:2 adapting:1 pre:1 anticipation:1 suggest:3 get:4 onto:1 selection:1 put:2 storage:1 influence:1 conventional:2 map:1 straightforward:1 attention:1 go:3 starting:1 exposure:8 hunger:1 thirst:1 rule:1 his:1 satiation:1 traditionally:1 variation:1 analogous:1 arranging:1 cs:4 target:1 imagine:1 construction:1 controlling:1 dickinson:9 substrate:2 us:9 sig:1 element:1 trend:1 particularly:2 corticocortical:1 updating:1 predicts:1 role:3 steven:1 russell:1 contemporary:1 substantial:2 environment:3 q1t:1 complexity:1 reward:22 cam:1 ultimately:1 trained:3 depend:1 grateful:1 exposed:1 predictive:6 patterson:1 completely:2 basis:1 montague:3 various:2 describe:2 london:1 effective:1 sejnowski:2 artificial:1 choosing:1 outcome:3 birth:2 sculpt:1 quite:2 whose:1 richer:1 solve:1 unconditioned:2 final:2 obviously:1 advantage:15 motivationally:3 net:1 ucl:1 reinforcer:1 rescorla:1 maximal:1 remainder:1 representational:1 competition:6 requirement:1 wired:11 produce:1 adam:1 executing:4 comparative:1 ac:1 pose:1 nearest:1 strong:1 c:8 involves:2 implies:1 come:3 correct:2 successor:7 premise:1 behaviour:2 really:1 repertoire:1 pessimistic:1 biological:1 extension:1 quinine:1 helping:1 wright:1 normal:1 presumably:1 lawrence:1 mapping:1 predict:1 cognition:2 circuitry:1 major:1 ventral:1 a2:8 determiner:1 pinette:1 sensitive:9 bridge:1 ofc:4 wl:1 pellet:6 mit:1 clearly:1 tegmental:1 always:2 rather:3 cr:1 tolman:1 eating:1 barto:3 basolateral:3 focus:1 improvement:1 mainly:2 contrast:1 psychiatry:1 suppression:1 sense:3 am:1 dayan:6 dependent:4 entire:1 kc:2 expand:1 interested:1 issue:1 overall:1 development:1 animal:20 emotion:1 construct:1 f3:1 vta:2 yu:1 future:10 report:2 stimulus:19 others:1 connectionist:1 sweet:1 sta:1 manipulated:1 delayed:1 phase:1 bw:3 saline:1 neuronlike:1 organization:1 evaluation:5 certainly:2 adjust:1 appetitive:6 mixture:1 light:1 pc:2 tj:2 dumb:1 occasioned:1 chain:6 necessary:1 lh:1 shorter:1 sweden:1 incomplete:2 re:9 theoretical:3 uncertain:1 psychological:5 instance:1 earlier:1 modeling:1 ar:1 devaluation:11 queen:1 cost:3 successful:1 seventh:1 erlbaum:1 too:1 motivating:1 straightforwardly:1 foraging:1 person:1 international:1 sensitivity:6 ie:2 interdisciplinary:1 standing:1 rewarding:1 together:1 quickly:1 thesis:3 choose:4 prefrontal:5 possibly:1 worse:1 berridge:4 cognitive:4 a1t:4 actively:1 account:3 protective:1 suggesting:1 potential:1 wk:1 includes:2 availability:1 baird:2 depends:1 view:8 responsibility:1 doing:1 start:3 sort:2 elicited:1 consummatory:1 contribution:1 square:1 air:1 kaufmann:1 characteristic:2 who:2 largely:1 dissociation:1 yes:1 plastic:5 garner:1 worth:5 drive:2 synaptic:1 colleague:2 obvious:1 dm:1 associated:7 irvine:1 intrinsically:2 murky:1 schedule:1 sophisticated:1 diet:1 response:6 improved:1 execute:2 though:1 evaluated:3 anderson:1 furthermore:1 stage:1 sketch:1 working:3 lack:1 willingness:1 perhaps:2 grows:1 effect:7 former:2 evolution:1 eg:2 ll:1 during:4 rat:4 muir:1 evident:2 vte:1 hungry:11 consideration:1 funding:1 functional:1 rl:18 empirically:1 salt:5 conditioning:17 attached:1 cohen:1 jl:1 discussed:1 interpretation:1 he:4 volume:1 neuropharmacology:1 cambridge:4 neuromodulatory:2 automatic:1 consistency:1 had:2 l3:2 specification:1 actor:3 cortex:7 access:2 base:2 add:1 showed:1 involvement:3 awarded:1 forcing:1 route:11 life:1 morgan:1 additional:1 greater:4 nigra:1 determine:1 paradigm:3 maximize:1 signal:3 dashed:5 relates:1 multiple:3 sham:1 hebbian:3 technical:1 match:1 england:2 academic:1 offer:2 believed:1 long:2 concerning:1 schizophrenia:1 paired:2 a1:4 controlled:4 prediction:11 involving:1 basic:1 controller:2 circumstance:1 expectation:2 heterogeneous:1 dopamine:9 iteration:4 adopting:1 cell:2 background:1 whereas:3 addition:1 interval:3 decreased:1 source:2 appropriately:1 extra:1 sr:12 subject:6 induced:1 undergo:1 recording:1 contrary:1 seem:1 conception:1 concerned:1 affect:1 psychology:8 gave:1 timesteps:1 approaching:1 competing:1 idea:2 prieditis:1 translates:1 absent:1 shift:1 whether:3 motivated:1 ultimate:2 peter:1 york:2 action:54 prepotent:1 useful:1 clear:1 involve:1 maybe:1 simplest:1 exist:1 dotted:2 estimated:1 neuroscience:4 sated:18 conceived:1 overly:1 instantly:1 incentive:10 affected:1 group:1 key:4 nevertheless:1 changing:1 neither:1 sum:7 compete:2 run:1 powerful:1 muli:1 place:1 strange:2 gonzalez:2 incompleteness:1 orbitofrontal:2 dorsolateral:1 drink:1 konorski:3 annual:1 activity:3 adapted:1 aspect:4 speed:1 pavlovian:19 dopaminergic:1 department:1 according:1 watt:1 kakade:1 modification:1 happens:2 sucrose:2 explained:1 pr:3 xo:12 depletion:1 turn:1 mechanism:2 fail:1 adopted:1 available:1 accorded:1 quarterly:1 appropriate:4 braver:1 jd:1 responding:2 remaining:1 include:1 worthless:1 tony:1 angela:1 opportunity:1 vigor:1 estes:1 build:1 classical:3 society:1 question:1 quantity:1 damage:1 primary:2 amongst:1 cw:1 link:2 separate:2 attentional:1 trivial:1 water:1 relationship:2 manipulanda:1 concomitant:1 providing:1 equivalently:1 difficult:1 executed:1 striatal:1 teach:1 debate:1 trace:1 negative:1 aversive:3 implementation:1 policy:6 perform:9 neuron:2 t:1 supporting:1 immediate:4 payoff:2 neurobiology:1 extended:2 looking:1 pair:1 connection:1 learned:10 able:1 suggested:1 corbit:1 lund:1 challenge:2 including:1 memory:3 explanation:1 critical:2 natural:5 force:3 treated:1 predicting:1 advanced:1 sodium:1 scheme:2 improve:1 declined:1 created:1 conventionally:1 mediated:1 sn:1 faced:1 prior:6 understanding:1 taste:2 acknowledgement:1 bee:1 review:1 fully:1 par:1 bear:1 abolished:1 suggestion:3 foundation:1 nucleus:5 contingency:4 offered:1 experi:1 principle:1 editor:3 charitable:1 critic:3 cd:1 course:1 bias:5 cortico:1 hedonic:1 curve:1 calculated:1 amygdala:9 transition:2 plain:1 cortical:1 chiba:1 world:3 made:1 reinforcement:6 expectancy:1 wouldn:1 collection:1 historical:1 far:1 schultz:1 adaptive:2 san:1 ec:1 compacta:1 mesencephalic:1 observable:1 uni:1 preferred:1 handbook:1 francisco:1 discriminative:2 gustatory:1 reviewed:1 learn:1 transfer:4 nature:1 ca:2 improving:1 interact:1 punishing:1 artificially:1 e20:1 da:1 main:3 decrement:1 motivation:20 edition:1 nothing:1 lesion:1 referred:1 creature:1 gatsby:3 ny:2 wiley:1 lc:1 experienced:2 sub:1 explicit:1 watkins:1 third:1 deprivation:3 learns:1 specific:3 physiological:3 consequent:1 ments:1 evidence:7 normalizing:1 intrinsic:5 dl:1 sequential:1 importance:1 barch:1 phd:3 gallagher:2 execution:2 conditioned:3 flavor:3 explore:1 unexpected:1 expressed:1 bo:1 holland:2 reflex:1 aa:1 chance:1 ma:1 shell:5 goal:9 careful:1 change:2 hard:11 included:1 determined:4 experimentally:1 generalisation:1 acting:2 correlational:1 total:3 secondary:2 partly:2 experimental:6 vrt:1 bernard:1 college:1 internal:2 tripartite:1 latter:2 dorsal:2 reactive:1 incorporate:1 phenomenon:1 mackintosh:1
1,150
2,049
The Noisy Euclidean Traveling Salesman Problem and Learning Mikio L. Braun, Joachim M. Buhmann braunm@cs.uni-bonn.de, jb@cs.uni-bonn.de Institute for Computer Science, Dept. III, University of Bonn R6merstraBe 164, 53117 Bonn, Germany Abstract We consider noisy Euclidean traveling salesman problems in the plane, which are random combinatorial problems with underlying structure. Gibbs sampling is used to compute average trajectories, which estimate the underlying structure common to all instances. This procedure requires identifying the exact relationship between permutations and tours. In a learning setting, the average trajectory is used as a model to construct solutions to new instances sampled from the same source. Experimental results show that the average trajectory can in fact estimate the underlying structure and that overfitting effects occur if the trajectory adapts too closely to a single instance. 1 Introduction The approach in combinatorial optimization is traditionally single-instance and worst-case-oriented. An algorithm is tested against the worst possible single instance. In reality, algorithms are often applied to a large number of related instances, the average-case performance being the measurement of interest. This constitutes a completely different problem: given a set of similar instances, construct solutions which are good on average. We call this kind of problem multiple-instances and average-case-oriented. Since the instances share some information, it might be expected that this problem is simpler than solving all instances separately, even for NP-hard problems. We will study the following example of a multiple-instance average-case problem, which is built from the Euclidean travelings salesman problem (TSP) in the plane. Consider a salesman who makes weekly trips. At the beginning of each week, the salesman has a new set of appointments for the week, for which he has to plan the shortest round-trip. The location of the appointments will not be completely random, because there are certain areas which have a higher probability of containing an appointment, for example cities or business districts within cities. Instead of solving the planning problem each week from scratch, a clever salesman will try to exploit the underlying density and have a rough trip pre-planned, which he will only adapt from week to week. An idealizing formulization of this setting is as follows. Fix the number of appointments n E N. Let Xl, ... , Xn E ]R2 and (J E 114. Then, the locations of the appointments for each week are given as samples from the normally distributed random vectors (i E {1, ... , n}) (1) The random vector (Xl, ... ,Xn ) will be called a scenario, sampled appointment locations (sampled) instance. The task consists in finding the permutation 7r E Sn which minimizes 7r I-t d7r (n)1f(l) + L~:ll d1f (i)1f(iH) , where dij := IIXi - Xj112' and Sn being the set of all bijective functions on the set {1, ... , n}. Typical examples are depicted in figure l(a)- (c). It turns out that the multiple-instance average-case setting is related to learning theory, especially to the theory of cost-based unsupervised learning. This relationship becomes clear if one considers the performance measure of interest. The algorithm takes a set of instances It, ... ,In as input and outputs a number of solutions Sl,???, Sn? It is then measured by the average performance (l/n) L~=l C(Sk, h), where C(s , I) denotes the cost of solution s on instance I. We now modify the performance measure as follows. Given a finite number of instances It, ... ,In, the algorithm has to construct a solution s' on a newly sampled instance I'. The performance is then measured by the expected cost E (C (s' ,I')). This can be interpreted as a learning task. The instances 11 , ... ,In are then the training data, E(C(s', I')) is the analogue of the expected risk or cost, and the set of solutions is identified with the hypothesis class in learning theory. In this paper, the setting presented in the previous paragraph is studied with the further restriction that only one training instance is present. From this training instance, an average solution is constructed, represented by a closed curve in the plane. This average trajectory is supposed to capture the essential structure of the underlying probability density, similar to the centroids in K-means clustering. Then, the average trajectory is used as a seed for a simple heuristic which constructs solutions on newly drawn instances. The average trajectories are computed by geometrically averaging tours which are drawn by a Gibbs sampler at finite temperature. This will be discussed in detail in sections 2 and 3. It turns out that the temperature acts as a scale or smoothing parameter. A few comments concerning the selection of this parameter are given in section 6. The technical content of our approach is reminiscent of the "elastic net" -approaches of Durbin and Willshaw (see [2], [5]) , but differs in many points. It is based on a completely different algorithmic approach using Gibbs sampling and a general technique for averaging tours. Our algorithm has polynomial complexity per Monte Carlo step and convergence is guaranteed by the usual bounds for Markov Chain Monte Carlo simulation and Gibbs sampling. Furthermore, the goal is not to provide a heuristic for computing the best solution, but to extract the relevant statistics of the Gibbs distribution at finite temperatures to generate the average trajectory, which will be used to compute solutions on future instances. 2 The Metropolis algorithm The Metropolis algorithm is a well-known algorithm which simulates a homogeneous Markov chain whose distribution converges to the Gibbs distribution. We assume that the reader is familiar with the concepts, we give here only a brief sketch of the relevant results and refer to [6], [3] for further details. Let M be a finite set and f: M -+ lit The Gibbs distribution at temperature T E Il4 is given by (m E M) 9T(m) := exp( - f(m)/T~ . exp( - f(m )/T) Lm/EM (2) The Metropolis algorithm works as follows. We start with any element m E M and set Xl +- m. For i ~ 2, apply a random local update m':= ?(Xi). Then set with probability min {I, exp( -(f(Xi) - f(m'))/T)} else (3) This scheme converges to the Gibbs distribution if certain conditions on ? are met. Furthermore, a L2-law of large numbers holds: For h: M --t ]R, ~ L:~=l h(X k ) --t L:mEM gT(m)h(m) in L2. For TSP, M = Sn and ? is the Lin-Kernighan two-change [4], which consists in choosing two indexes i, j at random and reversing the path between the appointments i and j. Note that the Lin-Kernighan two-change and its generalizations for neighborhood search are powerful heuristic in itself. 3 Averaging Tours Our goal is to compute the average trajectory, which should grasp the underlying structure common to all instances, with respect to the Gibbs measure at non-zero temperature T . The Metropolis algorithm produces a sequence of permutations 7rl, 7r2, ... with P{ 7rn = .} --t gT(.) for n --t 00. Since permutations cannot be added, we cannot simply compute the empirical means of 7rn . Instead, we map permutations to their corresponding trajectories. Definition 1 (trajectory) The trajectory of 7r E Sn given n points Xl, ... ,X n is a mapping r( 7r): {I, ... , n} --t ]R2 defined by r( 7r) (i) := X1C(i). The set of all trajectories (for all sets of n points) is denoted by Tn (this is the set of all mappings T {I , ... , n} --t ]R2 ). Addition of trajectories and multiplication with scalars can be defined pointwise. Then it is technically possible to compute L:~=l r(7rk). Unfortunately, this does not yield the desired results , since the relation between permutations and tours is not one-to-one. For example, the permutation obtained by starting the tour at a different city still corresponds to the same tour . We therefore need to define the addition of trajectories in a way which is independent of the choice of permutation (and therefore trajectory) to represent the tour. We will study the relationship between tours and permutations first in some detail, since we feel that the concepts introduced here might be generally useful for analyzing combinatorial optimization problems. t Definition 2 (tour and length of a tour) Let G = (V, E) be a complete (undirected) graph with V = {I, ... ,n} and E = (~). A subset tEE is called a tour iff It I = n, for every v E V, there exist exactly two el, e2 E t such that v E el and v E e2, and (V, t) is connected. Given a symmetric matrix (d ij ) of distances, the length of a tour t is defined by C(t) := L:{i,j} Et d ij . The tour corresponding to a permutation 7r E Sn is given by n-l t(7r) :={ {7r(I), 7r(n)}} U {{7r(i) ,7r(i + I)}}. U (4) i=l If t(7r) = t for a permutation 7r and a tour t, we say that 7r represents t. We call two permutations 7r, 7r' equivalent, if they represent the same tour and write 7r ,...., 7r'. Let [7r] denote the equivalence class of 7r as usual. Note that the length of a permutation is fully determined by its equivalence class. Therefore, ,...., describes the intrinsic symmetries of the TSP formulated as an optimization problem on Sn , denoted by TSP(Sn). We have to define the addition EB of trajectories such that the sum is independent of the representation. This means that for two tours h, t2 such that h is represented by 'lf1, 'If~ and t2 by 'lf2, 'If~ it holds that f('lf1) EB f('lf2) ~ f('lfD EB f('If~). The idea will be to normalize both summands before addition. We will first study the exact representation symmetry of TSP(Sn) ' The TSP(Sn) symmetry group Algebraically speaking, Sn is a group with concatenation of functions as multiplication, so we can characterize the equivalence classes of ~ by studying the set of operations on a permutation which map to the same equivalent class. We define a group action of Sn on itself by right translation ('If, 9 E Sn): " . " : Sn x Sn -+ Sn, g. 'If:= 'lfg- 1. (5) Note that any permutation in Sn can be mapped to another by an appropriate group action (namely 'If -+ 'If' by ('If,-l'lf) . 'If.), such that the group action of Sn on itself suffices to study the equivalence classes of ~. For certain 9 E Sn, it holds that t(g? 'If) = t('If). We want to determine the maximal set H t of elements which keeps t invariant. It even holds that H t is a subgroup of Sn: The identity is trivially in H t . Let g, h be t-invariant , then t((gh- 1) . 'If) = t(g ?(h- 1 . 'If)) = t(h- 1 . 'If) = t(h ?(h- 1 . 'If) = t( 'If). H t will be called the symmetry group of TSP(Sn) and it follows that ['If] = H t ? 'If :={h ? 'If I hE Hd. The shift u and reversal (2 are defined by (i E {I, ... , n} ) (.).__ {i + iz < n,n u z. 1 1 . = , (2(i) :=n + 1- i, (6) and set H :=((2, u), the group generated by u and (2. It holds that (this result is an easy consequence of (2(2 = id{l ,... ,n}, (2U = u- 1(2 and un = id{l ,... ,n}) H = {uk IkE {I, ... , n}} U {(2u k IkE {l, ... ,n}}. (7) The fundamental result is Theorem 1 Let t be the mapping which sends permutations to tours as defined in (4). Then, H t = H , where H t is the set of all t-invariant permutations and H is defined in (7). Proof: It is obvious that H ~ H t . Now, let h- 1 E H t . We are going to prove that t-invariant permutations are completely defined by their values on 1 and 2. Let hE H t and k:= h(l) . Then, h(2) = u(k) or h(2) = u - 1(k), because otherwise, h would give rise to a link {{'If(h(1),'If(h(2?}} 1. t('If) . For the same reason, h(3) must be mapped to u ?2(k). Since h must be bijective, h(3) =I- h(l) , so that the sign of the exponent must be the same as for h(2). In general, h(i) = u?(i- 1l(k). Now note that for i,k E {l , ... ,n } , u i(k) = uk(i) and therefore, h= { u k- 1 (2un-k if h(i) = ui-1 (k) ifh(i)=u- i+1(k)' D Adding trajectories We can now define equivalence for trajectories. First define a group action of Sn on Tn analogously to (5): the action of h E H t on "( E Tn is given by h ? "( := "( 0 h- 1 . Furthermore, we say that "( ~ 1} , if H t ? "( = H t ?1}. Our approach is motivated geometrically. We measure distances between trajectories as follows. Let d: ]R2 x ]R2 -+ Il4 be a metric. Then define h, 1} E Tn) dh,1}):= 2::=1 dh(k),1}(k). (8) Before adding two trajectories we will first choose equivalent representations "(', 1}' which minimize d( "(' , 1}'). Because of the results presented so far, searching through all equivalent trajectories is computationally tractable. Note that for h E H t , it holds that d( h . ,,(, h . rJ) = db, rJ) as h only reorders the summands. It follows that it suffices to change the representations only for one argument, since d(h? ,,(, i? rJ) = db, h - 1 i? rJ)? So the time complexity of one addition reduces to 2n computation of distances which involve n subtractions each. The normalizing action is defined by b, rJ E Tn) (9) n , 1J := argmin d( ,,(, n . rJ)? n EH t Assuming that the normalizing action is unique 1 , we can prove Theorem 2 Let ,,(, rJ be two trajectories, and n , 1J the unique normalizing action as defined in (9). Then, the operation "( EB rJ := "( + n , 1J . rJ (10) is representation invariant. Proof: Let "(I = g. ,,(, rJl = h? rJ for g, h E H t . We claim that n ,I1J1 The normalizing action is defined by n,I1J1 = gn' 1Jh-1. = argmin db /, n l . rJl) = argmin d(g . ,,(, nih? rJ) = argmin db , g-l n lh? rJ), n l EHt n l EH t n l EH t (11) by inserting g-l parallelly before both arguments in the last step. Since the normalizing action is unique, it follows that for the n l realizing the minimum it holds that g-ln l h = n , 1J and therefore n l = n , I1J1 = gn' 1Jh-1. Now, consider the sum which proves the representation independence. 0 The sum of more than two trajectories can be defined by normalizing everything with respect to the first summand, so that empirical sums EB~=l f(?ri) are now well-defined. t 4 Inferring Solutions on New Instances We transfer a trajectory to a new set of appointments Xl, .. . ,X n by computing the relaxed tour using the following finite-horizon adaption technique: First of all, passing times ti for all appointments are computed. We extend the domain of a trajectory "( from {I, ... , n} to the interval [1, n + 1) by linear interpolation. Then we define ti such that "((ti) is the earliest point with minimal distance between appointment Xi and the trajectory. The passing times can be calculated easily by simple geometric considerations. The permutation which sorts (ti)~l is the relaxed solution of"( to (Xi) . In a post-processing step, self-intersections are removed first. Then, segments of length w are optimized by exhaustive search. Let ?r be the relaxed solution. The path from ?rei) to ?r(i + w + 2) (index addition is modulo n) is replaced by the best alternative through the appointments ?r(i + 1), ... , ?r(i + w + 1). Iterate for all i E {I , . . . , n} until there is no further improvement. Since this procedure has time complexity w!n, it can only be done efficiently for small w. lOtherwise, perturb the locations of the appointments by infinitesimal changes. 5 Experiments For experiments, we used the following set-up: We took the 11.111-norm to determine the normalizing action. Typical sample-sizes for the Markov chain Monte Carlo integration were 1000 with 100 steps in between to decouple consecutive samples. Scenarios were modeled after eq. (1), where the Xi were chosen to form simple geometric shapes. Average trajectories for different temperatures are plotted in figures l(a)- (c). As the temperature decreases, the average trajectory converges to the trajectory of a single locally optimal tour. The graphs demonstrate that the temperature T acts as a smoothing parameter. To estimate the expected risk of an average trajectory, the post-processed relaxed (PPR) solutions were averaged over 100 new instances (see figure l(d)-(g)) in order to estimate the expected costs. The costs of the best solutions are good approximations, within 5% of the average minimum as determined by careful simulated annealing. An interesting effect occurs: the expected costs have their minimum at non-zero temperature. The corresponding trajectories are plotted in figure l(e),(f). They recover the structure of the scenario. In other words, average trajectories computed at temperatures which are too low, start to overfit to noise present only in the instance for which they were computed. So computation of the global optimum of a noisy combinatorial optimization problem might not be the right strategy, because the solutions might not reflect the underlying structure. Averaging over many suboptimal solutions provides much better statistics. 6 Selection of the Temperature The question remains how to select the optimal temperature. This problem is essentially the same as determining the correct model complexity in learning theory, and therefore no fully satisfying answer is readily available. The problem is nevertheless suited for the application of the heuristic provided by the empirical risk approximation (ERA) framework [1], which will be briefly sketched here. The main idea of ERA is to coarse-grain the set of hypotheses M by treating hypotheses as equivalent which are only slightly different. Hypotheses whose ?1 mutual distance (defined in a similar fashion as (8)) is smaller than the parameter "( E Il4 are considered statistically equivalent. Selecting a subset of solutions such that ?l -spheres of radius "( cover M results in the coarse-grained hypothesis set M,. VC-type large deviation bounds depending on the size of the coarse-grained hypothesis class can now be derived: p{ C2 (m"! ) - min C2 (m) mEM > 2c} :::; 21M"! 1 sup exp ( mEM., n(c am +c ( "()2 c - "( ) )' (13) am depending on the distribution. The bound weighs two competing effects. On the one hand, increasing "( introduces a systematic bias in the estimation. On the other hand, decreasing "( increases the cardinality of the hypothesis class. Given a confidence J > 0, the probability of being worse than c > 0 on a second instance and "( are linked. So an optimal coarsening "( can be determined. ERA then advocates to either sample from the ,,(-sphere around the empirical minimizer or average over these solutions. Now it is well known, that the Gibbs sampler is concentrated on solutions whose costs are below a certain threshold. Therefore, the ERA is suited for our approach. In the relating equation the log cardinality of the approximation set occurs, which is usually interpreted as micro canonical entropy. This relates back to statistical physics, the starting point of our whole approach. Now interpreting "( as energy, we can compute the stop temperature from the optimal T Using the well-known relation from statistical physics ~ee:t:~:: = T - 1 , we can derive a lower bound on the optimal temperature depending on variance estimates of the specific scenario given. 7 Conclusion In reality, optimization algorithms are often applied to many similar instances. We pointed out that this can be interpreted as a learning problem. The underlying structure of similar instances should be extracted and used in order reduce the computational complexity for computing solutions to related instances. Starting with the noisy Euclidean TSP, the construction of average tours is studied in this paper, which involves determining the exact relationship between permutation and tours, and identifying the intrinsic symmetries of the TSP. We hope that this technique might prove to be useful for other applications in the field of averaging over solutions of combinatorial problems. The average trajectories are able to capture the underlying structure common to all instances. A heuristic for constructing solutions on new instances is proposed. An empirical study of these procedures is conducted with results satisfying our expectations. In terms of learning theory, overfitting effects can be observed. This phenomenon points at a deep connection between combinatorial optimization problems with noise and learning theory, which might be bidirectional. On the one hand, we believe that noisy (in contrast to random) combinatorial optimization problems are dominant in reality. Robust algorithms could be built by first estimating the undistorted structure and then using this structure as a guideline for constructing solutions for single instances. On the other hand , hardness of efficient optimization might be linked to the inability to extract meaningful structure. These connections, which are subject of further studies, link statistical complexity to computational complexity. Acknowledgments The authors would like to thank Naftali Tishby, Scott Kirkpatrick and Michael Clausen for their helpful comments and discussions. References [1] J. M. Buhmann and M. Held. Model selection in clustering by uniform convergence bounds. Advances in Neural Information Processing Systems, 12:216- 222, 1999. [2] R. Durbin and D. Willshaw. An analogue approach to the travelling salesman problem using an elastic net method. Nature, 326:689- 691, 1987. [3] S. Kirkpatrick, C. D. Gelatt, and M. P. Vecchio Optimisation by simulated annealing. Science, 220:671- 680, 1983. [4] S. Lin and B. Kernighan. An effective heuristic algorithm for the traveling salesman problem. Operations Research, 21:498- 516, 1973. [5] P.D. Simic. Statistical mechanics as the underlying theory of "elastic" and "neural" optimizations. Network, 1:89-103, 1990. [6] G. Winkler. Image Analysis, Random fields and Dynamic Monte Carlo Methods, volume 27 of Application of Mathematics. Springer, Heidelberg, 1995. -sigma2 = O.03 i 17.7 &: 17.6 "j 17.5 o o f 17.4 <"Il T.,...,:0.15OO:Xl Lenglt. : 5.179571 0 o 17.3 o o temperatureT o.I H (d) e - si ma = O.025 ~ T.,...,: 0.212759 Lenglt. : 6.295844 11.5 o ~ &: 11 .45 "j 11.4 ~ ~ 11.35 o CD o temperatureT (f) n 5O"",11I>1S20_025 1 510 0_7654.2.() _742680 _2 31390 . 057 211.(l.Q1597 0 . 2 1 4 79 0.83 22 4 0 .58 33a1 ~ g Figure 1: (a) Average trajectories at different temperatures for n = 100 appointments on a circle with a 2 = 0.03. (b) Average trajectories at different temperatures, for multiple Gaussian sources, n = 50 and a 2 = 0.025. (c) The same for an instance with structure on two levels. (d) Average tour length of the post-processed relaxed (PPR) solutions for the circle instance plotted in (a). The PPR width was w = 5. The average fits to noise in the data if the temperature is too low, leading to overfitting phenomena. Note that the average best solution is :s: 16.5. (e) The average trajectory with the smallest average length of its PPR solutions in (d). (f) Average tour length as in (d). The average best solution is :s: 10.80. (g) Lowest temperature trajectory with small average PPR solution length in (f).
2049 |@word briefly:1 polynomial:1 norm:1 simulation:1 selecting:1 si:1 reminiscent:1 readily:1 must:3 grain:1 shape:1 treating:1 update:1 plane:3 beginning:1 realizing:1 coarse:3 provides:1 location:4 district:1 simpler:1 constructed:1 c2:2 consists:2 prove:3 advocate:1 paragraph:1 hardness:1 expected:6 planning:1 mechanic:1 decreasing:1 cardinality:2 increasing:1 becomes:1 provided:1 estimating:1 underlying:10 lowest:1 kind:1 interpreted:3 minimizes:1 argmin:4 finding:1 every:1 lfd:1 act:2 ti:4 braun:1 weekly:1 exactly:1 willshaw:2 uk:2 normally:1 before:3 local:1 modify:1 consequence:1 era:4 analyzing:1 id:2 path:2 interpolation:1 might:7 eb:5 studied:2 equivalence:5 statistically:1 averaged:1 unique:3 acknowledgment:1 differs:1 lf:1 procedure:3 area:1 empirical:5 pre:1 word:1 confidence:1 cannot:2 clever:1 selection:3 risk:3 restriction:1 equivalent:6 map:2 starting:3 identifying:2 hd:1 searching:1 traditionally:1 feel:1 construction:1 modulo:1 exact:3 homogeneous:1 hypothesis:7 element:2 satisfying:2 observed:1 capture:2 worst:2 connected:1 decrease:1 removed:1 complexity:7 ui:1 dynamic:1 solving:2 segment:1 technically:1 completely:4 easily:1 represented:2 d7r:1 effective:1 monte:4 choosing:1 neighborhood:1 exhaustive:1 whose:3 heuristic:6 say:2 otherwise:1 statistic:2 winkler:1 noisy:5 tsp:9 itself:3 sequence:1 net:2 took:1 maximal:1 inserting:1 relevant:2 iff:1 adapts:1 supposed:1 normalize:1 convergence:2 optimum:1 produce:1 converges:3 depending:3 derive:1 oo:1 undistorted:1 measured:2 ij:2 x1c:1 eq:1 c:2 involves:1 met:1 rei:1 radius:1 closely:1 correct:1 vc:1 everything:1 fix:1 generalization:1 suffices:2 hold:7 around:1 considered:1 exp:4 seed:1 algorithmic:1 week:6 mapping:3 lm:1 claim:1 consecutive:1 smallest:1 estimation:1 combinatorial:7 city:3 hope:1 rough:1 gaussian:1 idealizing:1 earliest:1 derived:1 joachim:1 improvement:1 contrast:1 centroid:1 am:2 helpful:1 el:2 relation:2 going:1 germany:1 sketched:1 denoted:2 exponent:1 plan:1 smoothing:2 integration:1 mutual:1 field:2 construct:4 clausen:1 sampling:3 represents:1 lit:1 ike:2 unsupervised:1 constitutes:1 future:1 jb:1 np:1 t2:2 summand:1 few:1 micro:1 oriented:2 familiar:1 replaced:1 interest:2 grasp:1 introduces:1 kirkpatrick:2 held:1 chain:3 lh:1 euclidean:4 desired:1 plotted:3 circle:2 weighs:1 minimal:1 instance:35 gn:2 planned:1 cover:1 cost:8 deviation:1 subset:2 tour:24 uniform:1 dij:1 conducted:1 too:3 tishby:1 characterize:1 answer:1 density:2 fundamental:1 systematic:1 physic:2 michael:1 analogously:1 vecchio:1 reflect:1 containing:1 choose:1 worse:1 leading:1 de:2 try:1 closed:1 linked:2 sup:1 start:2 sort:1 recover:1 minimize:1 il:1 variance:1 who:1 efficiently:1 yield:1 carlo:4 trajectory:38 definition:2 infinitesimal:1 against:1 energy:1 obvious:1 e2:2 proof:2 d1f:1 sampled:4 newly:2 stop:1 back:1 bidirectional:1 higher:1 done:1 furthermore:3 until:1 traveling:4 sketch:1 overfit:1 hand:4 kernighan:3 believe:1 effect:4 concept:2 symmetric:1 round:1 ll:1 self:1 width:1 naftali:1 simic:1 bijective:2 complete:1 demonstrate:1 tn:5 temperature:18 gh:1 interpreting:1 image:1 consideration:1 nih:1 common:3 rl:1 volume:1 discussed:1 he:4 extend:1 relating:1 measurement:1 refer:1 gibbs:10 trivially:1 mathematics:1 pointed:1 gt:2 summands:2 dominant:1 scenario:4 certain:4 tee:1 minimum:3 relaxed:5 subtraction:1 algebraically:1 shortest:1 determine:2 relates:1 multiple:4 rj:12 reduces:1 technical:1 adapt:1 sphere:2 lin:3 concerning:1 post:3 a1:1 essentially:1 metric:1 expectation:1 optimisation:1 represent:2 addition:6 want:1 separately:1 appointment:13 interval:1 else:1 annealing:2 source:2 sends:1 comment:2 subject:1 simulates:1 undirected:1 db:4 coarsening:1 call:2 ee:1 iii:1 easy:1 iterate:1 independence:1 fit:1 sigma2:1 identified:1 suboptimal:1 competing:1 reduce:1 idea:2 shift:1 motivated:1 speaking:1 passing:2 action:11 deep:1 generally:1 useful:2 clear:1 involve:1 iixi:1 locally:1 concentrated:1 processed:2 generate:1 sl:1 exist:1 canonical:1 sign:1 per:1 write:1 iz:1 group:8 nevertheless:1 threshold:1 drawn:2 graph:2 geometrically:2 sum:4 powerful:1 reader:1 bound:5 guaranteed:1 durbin:2 occur:1 ri:1 bonn:4 argument:2 min:2 lf2:2 describes:1 slightly:1 em:1 smaller:1 metropolis:4 invariant:5 computationally:1 ln:1 equation:1 remains:1 turn:2 tractable:1 reversal:1 travelling:1 salesman:8 studying:1 operation:3 available:1 apply:1 appropriate:1 gelatt:1 alternative:1 denotes:1 clustering:2 exploit:1 perturb:1 especially:1 prof:1 added:1 question:1 occurs:2 strategy:1 usual:2 rjl:2 distance:5 link:2 mapped:2 simulated:2 concatenation:1 thank:1 considers:1 reason:1 assuming:1 length:8 index:2 relationship:4 pointwise:1 modeled:1 unfortunately:1 eht:1 rise:1 guideline:1 markov:3 finite:5 rn:2 introduced:1 namely:1 trip:3 optimized:1 connection:2 subgroup:1 able:1 below:1 usually:1 scott:1 built:2 analogue:2 business:1 eh:3 buhmann:2 scheme:1 brief:1 extract:2 ppr:5 sn:22 geometric:2 l2:2 multiplication:2 determining:2 law:1 fully:2 permutation:20 interesting:1 lf1:2 share:1 cd:1 translation:1 last:1 bias:1 jh:2 institute:1 distributed:1 curve:1 calculated:1 xn:2 author:1 far:1 uni:2 keep:1 global:1 overfitting:3 mem:3 reorder:1 xi:5 search:2 un:2 sk:1 reality:3 scratch:1 nature:1 transfer:1 robust:1 elastic:3 symmetry:5 heidelberg:1 constructing:2 domain:1 main:1 whole:1 noise:3 mikio:1 fashion:1 inferring:1 xl:6 parallelly:1 grained:2 rk:1 theorem:2 specific:1 r2:6 normalizing:7 essential:1 intrinsic:2 ih:1 adding:2 horizon:1 suited:2 entropy:1 depicted:1 intersection:1 simply:1 scalar:1 springer:1 corresponds:1 minimizer:1 adaption:1 dh:2 extracted:1 ma:1 goal:2 formulated:1 identity:1 careful:1 content:1 hard:1 change:4 typical:2 determined:3 reversing:1 averaging:5 sampler:2 decouple:1 called:3 experimental:1 meaningful:1 select:1 il4:3 inability:1 dept:1 tested:1 phenomenon:2
1,151
205
516 Grossman The CHIR Algorithm for Feed Forward Networks with Binary Weights Tal Grossman Department of Electronics Weizmann Institute of Science Rehovot 76100 Israel ABSTRACT A new learning algorithm, Learning by Choice of Internal Represetations (CHIR), was recently introduced. Whereas many algorithms reduce the learning process to minimizing a cost function over the weights, our method treats the internal representations as the fundamental entities to be determined. The algorithm applies a search procedure in the space of internal representations, and a cooperative adaptation of the weights (e.g. by using the perceptron learning rule). Since the introduction of its basic, single output version, the CHIR algorithm was generalized to train any feed forward network of binary neurons. Here we present the generalised version of the CHIR algorithm, and further demonstrate its versatility by describing how it can be modified in order to train networks with binary (?1) weights. Preliminary tests of this binary version on the random teacher problem are also reported. I. INTRODUCTION Learning by Choice oflnternal Representations (CHIR) was recently introduced [1,11] as a training method for feed forward networks of binary units. Internal Representations are defined as the states taken by the hidden units of a network when patterns (e.g. from the training set) are presented to the input layer of the network. The CHIR algorithm views the internal representations associated with various inputs as the basic independent variables of the learning process. Once such representations are formed, the weights can be found by simple and local learning procedures such as the Percept ron Learning Rule (PLR) [2]. Hence the problem of learning becomes one of searching for proper internal representations, The CHIR Algorithm for Feed Forward Networks with Binary Weights rather than of minimizing a cost function by varying the values of weights, which is the approach used by back propagation (see, however [3],[4] where "back propagation of desired states" is described). This basic idea, of viewing the internal representations as the fundamental entities, has been used since by other groups [57]. Some of these works, and the main differences between them and our approach, are briefly disscussed in [11]. One important difference is that the CHIR algorithm, as well as another similar algorithm, the MRII [8], try to solve the learning problem for a fixed architecture, and are not guaranteed to converge. Two other algorithms [5,6] always find a solution, but at the price of increasing the network size during learning in a manner that resembles similar algorithms developed earlier [9,10]. Another approach [7] is to use an error minimizing algorithm which treat~ the internal representations as well as the weights as the relevant variables of the search space. To be more specific, consider first the single layer perceptron with its Perceptron Learning Rule (PLR) [2]. This simple network consists of N input (source) units j, and a single target unit i. This unit is a binary linear threshold unit, i.e. when the source units are set in anyone of Jl = 1, .. M patterns, i.e. Sj = the state of unit i, Si = ?1 is determined according to the rule ef, Si = sign(L WijSj + 0i) (1) . j Here Wij is the (unidirectional) weight assigned to the connection from unit j to ij 0i is a local bias. For each of the M input patterns, we require that the target unit (determined using (1)) will take a preassigned value Learning takes place in the course of a training session. Starting from any arbitrary initial guess for the weights, an input v is presented, resulting in the output taking some value Sf. Now modify every weight according to the rule er. (2) where TJ > 0 is a step size parameter (ej = 1 is used to modify the bias 0). Another input pattern is presented, and so on, until all inputs draw the correct output. The Perceptron convergence theorem states [2] that the PLR will find a solution (if one exists), in a finite number of steps. Nevetheless, one needs, for each unit, both the desired input and output states in order to apply the PLR. Consider now a two layer perceptron, with N input, H hidden and J{ output units (see Fig.1). The elements of the network are binary linear threshold units i, whose states Si = ?1 are determined according to (1). In a typical task for such a network, M specified output patterns, Sf'-,t,1J. = efut,lJ., are required in response to Jl 1, ... , M input patterns. If a solution is found, it first maps each input onto an internal representation generated on the hidden layer, which, in turn, produces the correct output. Now imagine that we are not supplied with the weights that solve the problem; however the correct internal representations are revealed. That is, we are given a table with M rows, one for each input. Every row has H bits ef'lJ. I for i = 1..H, specifying the state of the hidden layer obtained in response to input = 517 518 Grossman pattern 1'. One can now view each hidden-layer cell i as the target of the PLR, with the N inputs viewed as source. Given sufficient time, the PLR will converge to a set of weights Wii' connecting input unit j to hidden unit i, so that indeed the input-hidden association that appears in column i of our table will be realized. In order to obtain the correct output, we apply the PLR in a learning process that uses the hidden layer as source and each output unit as a target, so as to realize the correct output. In general, however, one is not supplied with a correct table of internal representations. Finding such a table is the goal of our approach . ... 0 Figure 1. A typical three layered feed forward network (two layered perceptron) with N input, H hidden and I( output units. The unidirectional weight Wij connects unit j to unit i. A layer index is implicitely included in each unit's index. During learning, the CHIR algorithm alternates between two phases: in one it generates the internal representations, and in the other it uses the updated representations in order to search for weights, using some single layer learning rule. This general scheme describes a large family of possible algorithms, that use different ways to change the internal representations. and update the weights. A simple algorithm based on this general scheme was introduced recently [1,11]. In section II we describe the multiple output version of CHIR [11]. In section III we present a way to modify the algorithm so it can train networks with binary weights, and the preliminary results of a few tests done on this new version. In the last section we shortly discuss our results and describe some future directions. The CHIR Algorithm for Feed Forward Networks with Binary Weights II. THE CHIR ALGORITHM The CHIR algorithm that we describe here implements the basic idea of learning by choice of internal representations by breaking the learning process into four distinct procedures that are repeated in a cyclic order: 1. SETINREP: Generate a table of internal representations {ef''''} by presenting each input pattern from the training set and recording the states of the hidden units, using Eq.(l), with the existing couplings Wij and 0i. 2. LEARN23: The current table of internal representations is used as the training set, the hidden layer cells are used as source, and each output as the target unit of the PLR. If weights Wij and 0i that produce the desired outputs are found, the problem has been solved. Otherwise stop after 123 learning sweeps, and keep the current weights, to use in CHANGE INREP. 3. CHANGE INREP: Generate a new table of internal representations, which reduces the error in the output : We present the table sequentially, row by row (pattern by pattern), to the hidden layer. If for pattern v the wrong output is obtained, the internal representation h 'lI is changed. e This is done simply by choosing (at random) a hidden unit i, and checking the effect of flipping the sign of on the total output error, i.e. the number of wrong bits. If the output error is not increased, the flip is accepted and the table of internal representations is changed accordingly. Otherwise the flip is rejected and we try another unit. When we have more than one output unit, it might happen that an error in one output unit can not be corrected without introducing an error in another unit. Therefore we allow only for a pre-specified number of attempted flips, lin, and go on to the next pattern even if the output error was not eliminated completely. This procedure ends with a "modified, "improved" table which is our next guess of internal representations. Note that this new table does not necessarily yield a totally correct output for all the patterns. In such a case, the learning process will go on even if this new table is perfectly realized by the next stage - LEARN12. e?'' ' 4. LEARN12: Present an input pattern; if the output is wrong, apply the PLR with the first layer serving as source, treating every hidden layer site separately as target. If input v does yield the correct output, we insert the current state of the hidden layer as the internal representation associated with pattern v, and no learning steps are taken. We sweep in this manner the training set, modifying weights Wij, (between input and hidden layer), hidden-layer thresholds Oi, and, as explained above, internal representations. If the network has achieved error-free performance for the entire training set, learning is completed. Otherwise, after lt2 training sweeps (or if the current internal representation is perfectly realized), abort the PLR stage, keeping the present values of Wij, Oi, and start SETINREP again. The idea in trying to learn the current internal representation even if it does not yield the perfect output is that it can serve as a better input for the next LEARN23 stage. That way, in each learning cycle the algorithm tries to improve the overall performance of the network. 519 520 Grossman This algorithm can be further generalized for multi-layered feed forward networks by applying the CHANGE INREP and LEARN12 procedures to each of the hidden layers, one by one, from the last to the first hidden layer. There are a few details that need to be added. a) The "iInpatience" parameters: lt2 and h3, which are rather arbitrary, are introduced to guarantee that the PLR stage is aborted if no solution is found, but they have to be large enough to allow the PLR to find a solution (if one exists) with sufficiently high probability. Similar considerations are valid for the lin parameter, the number of flip attempts allowed in the CHANGE INREP procedure. If this number is too small, the updated internal representations may not improve. If it is too large, the new internal representations might be too different from the previous ones, and therefore hard to learn. The optimal values depend, in general, on the problem and the network size. Our experience indicates, however, that once a "reasonable" range of values is found, performance is fairly insensitive to the precise choice. In addition, a simple rule of thumb can always be applied: "Whenever learning is getting hard, increase the parameters". A detailed study of this issue is reported in [11]. b) The Internal representations updating scheme: The CHANGE INREP procedure that is presented here (and studied in [11]) is probably the simplest and "most primitive" way to update the InRep table. The choice of the hidden units to be flipped is completely blind and relies only on the single bit of information about the improvement of the total output error. It may even happen that no change in the internal representaion is made, although such a change is needed. This procedure can certainly be made more efficient, e.g. by probing the fields induced on all the hidden units to be flipped and then choosing one (or more) of them by applying a "minimal disturbance" principle as in [8]. Nevertheless it was shown [11] that even this simple algorithm works quite well. c) The weights updating schemes: In our experiments we have used the simple PLR with a fixed increment (7] 1/2, .6.Wij ?1) for weight learning. It has the advantage of allowing the use of discrete (or integer) weights. Nevertheless, it is just a component that can be replaced by other, perhaps more sophisticated methods, in order to achieve, for example, better stability [12], or to take into account various constraints on the weights, e.g. binary weights [13]. In the following section we demonstrate how this can be done. = = III. THE CHIR ALGORITHM FOR BINARY WEIGHTS In this section we describe how the CHIR algorithm can be used in order to train feed forward networks with binary weights. According to this strong constraint, all the weights in the system (including the thresholds) can be either +1 or -1. The way to do it within the CHIR framework is simple: instead of applying the PLR (or any other single layer, real weights algorithm) for the updating of the weights, The CHIR Algorithm for Feed Forward Networks with Binary Weights we can use a binary perceptron learning rule. Several ways to solve the learning problem in the binary weight perceptron were suggested recently [13]. The one that we used in the experiments reported here is a modified version of the directed drift algorithm introduced by Venkatesh [13]. Like the standard PLR, the directed drift algorithm works on-line, namely, the patterns are presented one by one, the state of a unit i is calculated according to (1), and whenever an error occurs the incoming weights are updated. When there is an error it means that ,<0 ~'! hI! '-' e.n Wiie.r ' Namely, the field hi = Ej (induced by the current pattern is "wrong". If so, there must be some weights that pull it to the wrong direction. These are the weights for which erWii{r < o. er Here is the desired output of unit i for pattern v. The updating of the weights is done simply by flipping (i.e. Wii ~ -Wij ) at random k of these weights. The number of weights to be changed in each learning step, k, can be a prefixed parameter of the algorithm, or, as suggested by Venkatesh, can be decreased gradually during the learning process in a way similar to a cooling schedule (as in simulated annealing). What we do is to take k Ihl/2 + 1, making sure, like in relaxation algorithms, that just enough weights are flipped in order to obtain the desired target for the current pattern. This simple and local rule is now "plugged" into the Learn12 and Learn23 procedures instead of (2), and the initial weights are chosen to be + 1 or -1 at random. = We tested the binary version of CHIR on the "random teacher" problem. In this problem a "teacher network" is created by choosing a random set of +1/-1 weights for the given architecture. The training set is then created by presenting M input patterns to the network and recording the resulting output as the desired output patterns. Ip. what follows we took M 2N (exhaustive learning), and an N :N :1 architecture. = The "time" parameter that we use for measuring performance is the number of sweeps through the training set of M patterns ("epochs") needed in order to find the solution. Namely, how many times each pattern was presented to the network. In the experiments presented here, all possible input patterns were presented sequentially in a fixed order (within the perceptron learning sweeps). Therefore in each cycle of the algorithm there are 112 + h3 + 1 such sweeps. Note that according to our definition, a single sweep involves the updating of only one layer of weights or internal representations. for each network size, N, we created an ensemble of 50 independent runs, with different ranodom teachers and starting with a different random choice of initial weights. We calculate, as a performance measure, the following quantities: a. The median number of sweeps, t m . b. The "inverse average rate", T, as defined by Tesauro and Janssen in [14]. 521 522 Grossman c. The success rate, S, i.e. the fraction of runs in which the algorithm finds a solution in less than the maximal number of training cycles [max specified. The results,with the typical parameters, for N=3,4,5,6, are given in Table 1. Table 1. The Random Teacher problem with N:N:l architecture. N lt2 123 lin [max tm T S 3 4 5 6 20 25 40 70 10 10 15 40 5 7 9 11 20 60 300 900 14 87 430 15000 9 37 60 1100 1.00 1.00 1.00 0.71 As mentioned before, these are only preliminary results. No attempt was made to to optimize the learning parameters. IV. DISCUSSION We presented a generalized version of the CHIR algorithm that is capable of training networks with multiple outputs and hidden layers. A way to modify the basic alf$ortihm so it can be applied to networks with binary weights was also explained and tested. The potential importance of such networks, e.g. in hardware implementation, makes this modified version particularly interesting. An appealing feature of the CHIR algorithm is the fact that it does not use any kind of "global control", that manipulates the internal representations (as is used for example in [5,6]). The mechanism by which the internal representations are changed is local in the sense that the change is done for each unit and each pattern without conveying any information from other units or patterns (representations). Moreover, the feedback from the "teacher" to the system is only a single bit quantity, namely, whether the output is getting worse or not (in contrast to BP, for example, where one informs each and every output unit about its individual error). Other advantages of our algorithm are the simplicity of the calculations, the need for only integer, or even binary weights and binary units, and the good performance. It should be mentioned again that the CHIR training sweep involves much less computations than that of back-propagation. The price is the extra memory of M H bits that is needed during the learning process in order to store the internal representations of all M training patterns. This feature is biologically implausible and may be practically limiting. We are developing a method that does not require such memory. The learning method that is currently studied for that purpose [15], is related to the MRII rule, that was recently presented by Widrow and Winter in [8]. It seems that further research will be needed in order to study the practical differences and the relative advantages of the CHIR and the MRII algorithms. The eHIR Algorithm for Feed Forward Networks with Binary Weights Acknowledgements: I am gratefull to Prof. Eytan Domany for many useful suggestions and comments. This research was partially supported by a grant from Minerva. References [1] Grossman T., Meir R. and Domany E., Complex Systems 2, 555 (1989). See also in D. Touretzky (ed.), Advances in Neural Information Processing Systems 1, (Morgan Kaufmann, San Mateo 1989). [2] Minsky M. and Papert S. 1988, Perceptrons (MIT, Cambridge); Rosenblatt F. Principles of neurodynamics (Spartan, New York, 1962). [3] Plaut D.C., Nowlan S.J., and Hinton G.E., Tech.Report CMU-CS-86-126, Carnegie-Mellon University (1986). [4] Le Cun Y., Proc. Cognitiva 85, 593 (1985). [5] Rujan P. and Marchand M., in the Proc. of the First International Joint Conference Neural Networks - Washington D. C. 1989, Vol.lI, pp. 105. and to appear in Complex Systems. [6] Mezard M. and Nadal J.P., J.Phys.A. 22, 2191 (1989). [7] Krogh A., Thorbergsson G.1. and Hertz J.A., in these Proceedings. R. Rohwer, to apear in the Proc. of DANIP, GMD Bonn, April 1989, J. Kinderman and A. Linden eds ; Saad D. and Merom E., preprint (1989). [8] Widrow B. and Winter R., Computer 21, No.3, 25 (1988). [9] See e.g. Cameron S.H., IEEE TEC EC-13,299 (1964) ; Hopcroft J.E. and Mattson R.L., IEEE, TEC EC-14, 552 (1965). [10] Honavar V. and' Uhr L. in the Proc. of the 1988 Connectionist Models Summer School, Touretzky D., Hinton G . and Sejnowski T. eds. (Morgan Kaufmann, San Mateo, 1988). [11] Grossman T., to be published in Complex Systems (1990). [12] Krauth W. and Mezard M., J.Phys.A, 20, L745 (1988). [13] Venkatesh S., preprint (1989) ; Amaldi E. and Nicolis S., J.Phys.France 50, 2333 (1989). Kohler H., Diederich S., Kinzel W. and Opper M., preprint (1989). [14] Tesauro G. and Janssen H., Complex Systems 2, 39 (1988). [15] Nabutovski D., unpublished. 523
205 |@word briefly:1 version:9 seems:1 electronics:1 cyclic:1 initial:3 existing:1 current:7 nowlan:1 si:3 must:1 realize:1 happen:2 treating:1 update:2 guess:2 accordingly:1 plaut:1 ron:1 wijsj:1 consists:1 manner:2 indeed:1 aborted:1 multi:1 increasing:1 becomes:1 totally:1 moreover:1 israel:1 what:2 kind:1 nadal:1 developed:1 finding:1 guarantee:1 every:4 wrong:5 control:1 unit:35 grant:1 appear:1 generalised:1 before:1 local:4 treat:2 modify:4 preassigned:1 might:2 resembles:1 studied:2 mateo:2 specifying:1 kinderman:1 range:1 weizmann:1 directed:2 practical:1 implement:1 procedure:9 pre:1 onto:1 layered:3 applying:3 optimize:1 map:1 go:2 primitive:1 starting:2 simplicity:1 manipulates:1 rule:10 pull:1 stability:1 searching:1 increment:1 limiting:1 updated:3 target:7 imagine:1 us:2 element:1 particularly:1 updating:5 cooling:1 cooperative:1 preprint:3 solved:1 calculate:1 cycle:3 chir:22 mentioned:2 depend:1 serve:1 completely:2 joint:1 hopcroft:1 various:2 train:4 distinct:1 describe:4 sejnowski:1 spartan:1 tec:2 choosing:3 exhaustive:1 whose:1 quite:1 solve:3 otherwise:3 ip:1 advantage:3 took:1 maximal:1 adaptation:1 relevant:1 achieve:1 setinrep:2 getting:2 convergence:1 produce:2 perfect:1 coupling:1 informs:1 widrow:2 ij:1 school:1 h3:2 eq:1 krogh:1 strong:1 c:1 involves:2 direction:2 correct:8 modifying:1 uhr:1 viewing:1 require:2 preliminary:3 insert:1 practically:1 sufficiently:1 alf:1 purpose:1 proc:4 currently:1 mit:1 always:2 modified:4 rather:2 ej:2 varying:1 improvement:1 indicates:1 tech:1 contrast:1 sense:1 am:1 lj:2 entire:1 hidden:22 wij:8 france:1 overall:1 issue:1 fairly:1 field:2 once:2 washington:1 eliminated:1 flipped:3 amaldi:1 future:1 report:1 connectionist:1 few:2 winter:2 individual:1 replaced:1 phase:1 connects:1 minsky:1 versatility:1 attempt:2 certainly:1 tj:1 capable:1 experience:1 iv:1 plugged:1 desired:6 minimal:1 increased:1 column:1 earlier:1 measuring:1 cost:2 introducing:1 too:3 reported:3 teacher:6 fundamental:2 international:1 connecting:1 again:2 worse:1 grossman:7 li:2 account:1 potential:1 inrep:6 blind:1 view:2 try:3 start:1 unidirectional:2 formed:1 oi:2 kaufmann:2 percept:1 ensemble:1 yield:3 conveying:1 thumb:1 published:1 implausible:1 phys:3 touretzky:2 whenever:2 ed:3 diederich:1 definition:1 rohwer:1 pp:1 associated:2 stop:1 ihl:1 schedule:1 sophisticated:1 back:3 appears:1 feed:10 response:2 improved:1 april:1 done:5 learn23:3 rejected:1 stage:4 just:2 until:1 propagation:3 abort:1 perhaps:1 effect:1 hence:1 assigned:1 during:4 generalized:3 trying:1 presenting:2 demonstrate:2 representaion:1 consideration:1 ef:3 recently:5 krauth:1 kinzel:1 insensitive:1 jl:2 association:1 mellon:1 cambridge:1 session:1 tesauro:2 store:1 binary:21 success:1 morgan:2 converge:2 ii:2 multiple:2 reduces:1 calculation:1 lin:3 cameron:1 basic:5 cmu:1 minerva:1 achieved:1 cell:2 whereas:1 addition:1 separately:1 decreased:1 annealing:1 median:1 source:6 extra:1 saad:1 cognitiva:1 probably:1 comment:1 sure:1 recording:2 induced:2 integer:2 revealed:1 iii:2 enough:2 architecture:4 perfectly:2 plr:15 reduce:1 idea:3 tm:1 domany:2 whether:1 york:1 useful:1 detailed:1 hardware:1 simplest:1 gmd:1 generate:2 supplied:2 meir:1 sign:2 serving:1 rosenblatt:1 rehovot:1 discrete:1 carnegie:1 vol:1 group:1 four:1 threshold:4 nevertheless:2 relaxation:1 fraction:1 run:2 inverse:1 place:1 family:1 reasonable:1 draw:1 bit:5 layer:21 hi:2 guaranteed:1 summer:1 marchand:1 constraint:2 bp:1 implicitely:1 tal:1 generates:1 bonn:1 anyone:1 department:1 developing:1 according:6 alternate:1 honavar:1 hertz:1 describes:1 appealing:1 cun:1 making:1 biologically:1 lt2:3 explained:2 gradually:1 taken:2 describing:1 turn:1 discus:1 mechanism:1 needed:4 flip:4 prefixed:1 end:1 wii:2 apply:3 shortly:1 completed:1 prof:1 sweep:9 added:1 realized:3 flipping:2 occurs:1 quantity:2 simulated:1 entity:2 index:2 minimizing:3 implementation:1 proper:1 allowing:1 neuron:1 finite:1 hinton:2 precise:1 arbitrary:2 drift:2 introduced:5 venkatesh:3 namely:4 required:1 specified:3 unpublished:1 connection:1 suggested:2 pattern:27 including:1 max:2 memory:2 disturbance:1 scheme:4 improve:2 created:3 epoch:1 acknowledgement:1 checking:1 relative:1 interesting:1 suggestion:1 sufficient:1 principle:2 row:4 course:1 changed:4 supported:1 last:2 free:1 keeping:1 bias:2 allow:2 perceptron:9 institute:1 taking:1 feedback:1 calculated:1 opper:1 valid:1 forward:10 made:3 san:2 ec:2 sj:1 keep:1 global:1 sequentially:2 incoming:1 search:3 table:15 neurodynamics:1 learn:2 necessarily:1 complex:4 rujan:1 main:1 repeated:1 allowed:1 fig:1 site:1 probing:1 papert:1 mezard:2 sf:2 breaking:1 theorem:1 specific:1 er:2 linden:1 exists:2 janssen:2 importance:1 simply:2 partially:1 applies:1 thorbergsson:1 relies:1 viewed:1 goal:1 price:2 change:9 hard:2 included:1 determined:4 typical:3 corrected:1 total:2 accepted:1 eytan:1 attempted:1 perceptrons:1 internal:32 learn12:4 kohler:1 tested:2
1,152
2,050
Linear Time Inference in Hierarchical HMMs Kevin P. Murphy and Mark A. Paskin Computer Science Department University of California Berkeley, CA 94720-1776 murphyk,paskin @cs.berkeley.edu  Abstract The hierarchical hidden Markov model (HHMM) is a generalization of the hidden Markov model (HMM) that models sequences with structure at many length/time scales [FST98]. Unfortunately, the original infertime, where is ence algorithm is rather complicated, and takes the length of the sequence, making it impractical for many domains. In this paper, we show how HHMMs are a special kind of dynamic Bayesian network (DBN), and thereby derive a much simpler inference algorithm, which only takes time. Furthermore, by drawing the connection between HHMMs and DBNs, we enable the application of many standard approximation techniques to further speed up inference.    1 Introduction The Hierarchical HMM [FST98] is an extension of the HMM that is designed to model domains with hierarchical structure, e.g., natural language, XML, DNA sequences [HIM 00], handwriting [FST98], plan recognition [BVW00], visual action recogntion [IB00, ME01, Hoe01], and spatial navigation [TRM01, BVW01]. HHMMs are less expressive than stochastic context free grammars (SCFGs), since they only allows hierarchies of bounded depth, but they are more efficient and easier to learn. Unfortunately, the original inference algorithm described in [FST98] is somewhat complicated, and takes time, where is the length of the sequence, is the depth of the hierarchy, and is the (maximum) number of states at each level of the hierarchy. In this paper, we show how to represent an HHMM as a dynamic Bayesian network (DBN), and thereby derive a much simpler and faster inference algorithm, which takes at most time; empirically, we find it takes only time using the junction tree algorithm. Furthermore, by drawing the connection between HHMMs and DBNs, we enable the application of approximate inference techniques such as belief propagation, which can perform inference in time.         !         " By inference, we mean offline smoothing, i.e., conditioning on a fixed-length observation se- quence. This is needed as a subroutine for EM. Once the model has been learned, it will typically be used for online inference (filtering). end 0 2 3 a end 1 end 4 5 b 6 c 8 x 9 7 end d end y Figure 1: A 3-level hierarchical automaton representing the regular expression     . Solid lines represent horizontal transitions, dotted lines represent vertical transitions. Letters below a production state represent the symbol that is emitted. The unnumbered root node is considered level 0, and could be omitted if we fully interconnected states 0 and 1.   We will describe HHMMs in Section 2, and the original inference algorithm in Section 3. The main contribution of the paper is in Section 4, where we show how to represent an HHMM as a DBN. In Section 5, we discuss how to do efficient inference in this DBN, and in Section 6, we discuss related work. In the full version of this paper, we discuss how to do parameter and structure learning using EM. 2 Hierarchical HMMs HHMMs are like HMMs except the states of the stochastic automaton can emit single observations or strings of observations. (For simplicity of exposition, we shall assume all observations are discrete symbols, but HHMMs can easily be generalized to handle continuous observations, as we discuss in Section 4.1.) Those that emit single symbols are called ?production states?, and those that emit strings are termed ?abstract states?. The strings emitted by abstract states are themselves governed by sub-HHMMs, which can be called recursively. When the sub-HHMM is finished, control is returned to wherever it was called from; the calling context is memorized using a depth-limited stack. We illustrate the generative process with Figure 1, which shows the state transition diagram of an example HHMM which models the regular expression       . We start in the root state, and make a ?vertical transition? to one of its children, say state 0. From here, we make another vertical transition to state 2. Since state 2 is a production state, it emits ?a? and then makes a ?horizontal transition? to state 3. State 3 calls its sub-HMM, which emits x?s and y?s until it enters its end state; then control is returned to the calling state, in this case state 3. We then make a horizontal transition to state 4, emit ?b?, and enter the end state, thereby returning control to state 0. Finally, from state 0, we return control to the root, and optionally start again.   Any HHMM can be converted to an HMM by creating a state for every possible legal stack configuration !  . If the HHMM transition diagram is a tree, there will be one HMM state for every HHMM production state. If the HHMM transition diagram has shared substructure (such as the sub-expression  ), this structure must be duplicated in the HMM, generally resulting in a larger model. It is the ability to reuse sub-models in different con-    texts that makes HHMMs more powerful than standard HMMs. In particular, the parameters of such shared sub-models only need to be learned once. (Given segmented data, we can train the sub-HMMs separately, and then ?glue them together?, but it is also possible to train the HHMM on unsegmented data; see the full version of this paper for details.) 3 Cubic-time inference  The inference algorithm for HHMMs presented in [FST98] runs in time and is based on the Inside-Outside algorithm [LY90], an exact inference algorithm for stochastic context-free grammars (SCFGs) which we now describe. In an SCFG, sequences of observations are generated by a set of stochastic production into either rules. Each production rule stochastically rewrites a non-terminal symbol  ) or a pair of nonterminal symbols ( a symbol of the alphabet ( ). Observation strings are generated by starting with the distinguished ?start? nonterminal , and continually re-writing all non-terminals using stochastic production rules until, finally, only symbols of the alphabet remain.                   ! ! The Inside-Outside algorithm computes , where ! !    ! ! ! is a subsequence. This can then be used to compute the expected sufficient statistics needed by the EM algorithm to learn the parameters of the model. If there are non-terminals in the grammar and the training sequence is of length , then the time. To see why, note that we must compute Inside-Outside algorithm requires ! ! for all end points and , and for all midpoints such that   ! generates ! ! and generates ! ? the three degrees for freedom ,   and gives rise to the term. The term arises because we must consider all , and .                       !       The inference algorithm for HHMMs presented in [FST98] is based upon a straightforward adaptation of the Inside-Outside algorithm. The algorithm computes in state at time  ! ! by assuming that sub-state generates ! ! , that a   ! ! transition to state occurs, and that generates . Hence the algorithm  takes time, where is the total number of states.  "$#      % " # " #       We can always ?flatten? an HHMM into a regular HMM and hence do inference in . Unfortunately, this flat model cannot represent the hierarchical structure, yet alone learn it. In the next section, we show how to represent the HHMM as a DBN, and thereby get the best of both worlds: low time complexity without losing hierarchical structure.    4 Representing the HHMM as a DBN We can represent the HHMM as a dynamic Bayesian network (DBN) as shown in Figure 2. (We assume for simplicity that all production states are at the bottom of the hierarchy; this restriction is lifted in the full version of this paper.) The state of the HMM at level  and time is represented by ! . The state of the whole HHMM is encoded by the vector ! ! ! ; intuitively, this encodes the contents of the stack, that specifies the complete ?path? to take from the root to the leaf state.  '     (  ) # ! &# ) #+  * ) #  ,* is an indicator variable that is ?on? if the HMM at level  and time has just ?finished? (i.e., is about to enter an end state), otherwise it is off. Note that if ! , then ! )  )    )   )   )    )       )     *  )   )        Figure 2: An HHMM represented as a DBN.                is the state at time , level ; if the . Shaded nodes are observed; HMM at level has finished (entered its exit state), otherwise the remaining nodes are hidden. We may optionally clamp , where is the length of the observation sequence, to ensure all models have finished by the end of the sequence. (A similar trick was used in [Zwe97].)  )  nodes that are ?off? represents the effective height for all   ; hence the number of of the ?context stack?, i.e., which level of the hierarchy we are currently on.  ) The downward going arcs between the variables represent the fact that a state ?calls? a sub-state. The upward going arcs between the variables enforce the fact that a higherlevel HMM can only change state when the lower-level one is finished. This ensures proper nesting of the parse trees, and is the key difference between an HHMM and a hidden Markov decision tree [JGS96]. We will define the conditional probability distributions (CPDs) of each of the node types below, which will complete the definition of the model. We consider the bottom, middle and top layers of the hierarchy separately (since they have different local topology), as well as the first, middle and last time slices. 4.1 Definition of the CPDs   follows a Markov chain with parameters Consider the bottom level of the hierarchy. determined by its position in the automaton, which is encoded by the vector of higher-up ! ! state variables !  , which we will represent by the integer . When      (    %   If the topology is sparse, this distribution will be 0 for many values of  . This will be discussed in Section 4.2. )    enters its end state, it will ?turn on? , to mean it is finished; this will be a signal that higher-level HMMs can now change state. In addition, it will be a signal that the next value of should be drawn from its prior distribution (representing a vertical transition), instead of its transition matrix (representing a horizontal transition). Formally, we can write this as follows:          )        %       if           if * where we have assumed    end.  is the transition matrix for level  given that the parent variables are in state % , and   is just a rescaled version of  .  Similarly,  is the initial distribution for level  given that the parent variables are in state % . The equation ) for  is simply  )   *     %          end  Now consider the intermediate levels. As before,  # follows a Markov chain with param) eters determined by  #  , and # specifies whether we should use the) transition matrix or the prior. The difference is that we now also get a signal from below, # , specifying ! !  ! ! !  !   !  whether the sub-model has finished or not; if it has, we are free to change state, otherwise we must remain in the same state. Formally, we can write this as follows:         if  ) )               *      # #  #  %  #  #  # #   ifif  * and and   * ) # should ?turn on? only if  # is ?allowed? to enter a final state, the probability of which depends on the current context  #  . Formally, we can write this as follows:   ) #  *  #    #   %  ) #    #    end ifif  *  !  ! ! !  !     !  ! !  !      The top level differs from the intermediate levels in that the node has no parent to specify which distribution to use. The equations are the same as above, except we eliminate the conditioning on !  . (Equivalently, we can imagine a dummy top layer HMM, which is always in state 1: ! . This is often how HHMMs are represented, so that this top-level state is the root of the overall parse tree, as in Figure 1.)  #   % *           .  for the top ' If the observations are discrete symbols, we may represent    as a multinomial (i.e., using a table), or by using any of the more parsimonious representations discussed in Section' 4.2. If the observations are real-valued vectors, we can use a Gaussian for each value of  , or a mixture of a smaller number of Gaussians, as in [GJ97].  Unlike the automaton representation, the DBN never actually enters an end state (i.e.,   can never taken on the value ?end?), because if it did, it would not be able to emit the symbol   . Instead, on, and then enters a new (non-terminal) state at time  . This means   causes   andto turn that HHMM are not identical, but satisfy the following   "the!$# &DBN % ('*)    "+   transition "!$# ,! wherematrices  represents   relation: the automaton transition matrix, represents  .-0 /$1   "! end is the probability of terminating from state  . the DBN transition matrix, and )   The equations holds because the probability of each transition in the DBN gets multiplied  ; thishorizontal by the probability that    , which is 2'3)   product should match the original probability.   #   #   %  #  The CPDs for the nodes in the first slice are as follows:  level and , for   ! !             It is easy to see that the new matrix is also stochastic, as required.  ! 4.2 Parsimonious representations of the CPDs   # &#    #   % !  ! !  The number of parameters needed to represent as a multinomial is . If the state-transition diagram of the hierarchical automaton is sparse, many of the entries in this table will be 0. However, when we are learning a model, we do not know the structure of the state-transition diagram, and must therefore adopt a representation with fewer parameters. There are at least three possibilities: decision trees [BFGK96], softmax !  ! !  as a mixture of smaller transition matrices nodes, or representing at different depths c.f. [SJ99]. See the full version of this paper for details.   #  &#  #    #   % 5 Linear-time inference            parents We define inference to be computing   for all sets of nodes  in the DBN. These ?family? marginals are needed by EM. The simplest way to do this is to merge all the hidden nodes in each slice into a single ?mega node?,  ! , with  possible values. (The term arises from the binary nodes.) We can then apply the forwards-backwards algorithm for HMMs, which takes  time.  )    Unfortunately, converting the DBN to an HMM in this way will not be tractable for reasonably large or . (Even storing the  transition matrix is likely to consume too much space.) Fortunately, we can do better by exploiting the structure of the model. In [Mur01], we present a way of applying the junction tree (jtree) algorithm to variable-length DBNs; we give a brief sketch here. The algorithm works by performing a forwards-backwards sweep through a chain of jtrees. Each jtree is formed from a ? -slice DBN?; this is a DBN that contains all the nodes in slice 1 but only the interface nodes from slice 2. The interface nodes are those nodes in slice 2 that have an incoming temporal arc, plus parents of nodes that have incoming temporal arcs. In the case of an HHMM, the interface is all the nodes.   *   The cost of doing inference in each jtree depends on the sizes of the cliques. Minimizing the maximal clique size is NP-hard, so we used a standard one-step look-ahead (greedy) algorithm [Kja90]. The resulting cliques are hard to interpret, but we can still analyze the complexity. Let  be the number of nodes in clique  , let  be the number of nodes, and let  be the number of cliques. Then the cost of inference in a jtree is proportional to    )                 !      "$#&%('  !     *  *                         )*        )+ #&%('       Empirically we find that, for a wide2 range of ,  , ,-/.  10 2 3 4 and ,5-/. 6 73 4 . Hence a crude upper bound on the cost of in"0  ference in each jtree is , yielding an overall time and space complexity of . We remind readers that the original algorithm has time complexity, since there can be up to states in the HHMM. The advantage of the new algorithm in practice is clearly illustrated in Figure 3.          We can reduce the time (and space) complexity from to by using approximate DBN inference techniques such as the ?factored frontier (FF) algorithm? [MW01], which is equivalent to applying ?loopy belief propagation? to the DBN using a left-right scheduling of the messages. (It is still exponential in because of the high fan-in of the nodes.) We can get a further speedup by using a mixture representation of the CPDs  running time (seconds) vs sequence length 40 35 30 25 20 15 10 5 0 10 linear cubic 15 20 25 30 35 40  ,   . Figure 3: Running time vs. sequence length. Both algorithms were implemented in Matlab. The HHMM has (see Section 4.2). In this case, we can exploit the form of the CPD to compute the required messages efficiently [Mur99], bringing the overall complexity down to .      We remark that all of the above algorithms can also be used for online filtering. In addition, by replacing the sum operator with max, we can do Viterbi segmentation in the usual way. 6 Related work ) Hidden Markov decision trees (HMDT) [JGS96] are DBNs with a structure similar to Figure 2, but they lack the nodes and the upward going arcs; hence they are not able to represent the call-return semantics of the HHMM. Embedded HMMs [NI00] are a special case of HHMMs in which the ending ?time? of the sub-HMMs is known in advance (e.g., the sub-HMM models exactly one row of pixels). ([Hoe01] calls these models ?hierarchical mixture of Markov chains?.) A variable-duration HMM [Rab89] is a special case of a 2-level HHMM, where the bottom level counts how long we have been in a certain state; when the counter expires, the node turns on, and the parent can change state. ) [BVW00] describes the ?Abstract HMM? (AHMM), which is very closely related to HHMMs. These authors are interested in inferring what abstract policy an agent is following by observing its effects in the world. An AHMM is equivalent to an HHMM if we consider ! to represent the (abstract) policy being followed at level  and time ; ! represents the concrete action, which causes the observation. We also need to add a hidden global nodes. state variable ! , which is a parent of the ! node, all the ! nodes and all the ! ( ! is hidden to us as observers, but not to the agent performing the actions.) [BVW00] consider abstract policies of the ?options? kind [SPS99], which is equivalent to assuming that there are no horizontal transitions. (HAMs [PR97] generalize this by allowing horizontal transitions (i.e., internal state) within a controller.) In addition, they assume that ! only depends on its immediate parent, ! , but not its whole context, !  , so the nodes become connected by a chain. This enables them to use Rao-Blackwellized particle filtering for approximate online inference: conditioned on the nodes, the distribution over the nodes can be represented as a product of marginals, so they can be efficiently marginalized out.  #    # )    )   #  #  Acknowledgements I would like to thank Dr Christopher Schlick for giving me his Matlab implementation of the algorithm, which was used to create part of Figure 3.  References [BFGK96] C. Boutilier, N. Friedman, M. Goldszmidt, and D. Koller. Context-Specific Independence in Bayesian Networks. In UAI, 1996. [BVW00] H. Bui, S. Venkatesh, and G. West. On the recognition of abstract Markov policies. In AAAI, 2000. [BVW01] H. Bui, S. Venkatesh, and G. West. Tracking and surveillance in wide-area spatial environments using the Abstract Hidden Markov Model. Intl. J. of Pattern Rec. and AI, 2001. [FST98] Shai Fine, Yoram Singer, and Naftali Tishby. The hierarchical Hidden Markov Model: Analysis and applications. Machine Learning, 32:41, 1998. [GJ97] Z. Ghahramani and M. Jordan. Factorial hidden Markov models. Machine Learning, 29:245?273, 1997. [HIM 00] M. Hu, C. Ingram, M.Sirski, C. Pal, S. Swamy, and C. Patten. A Hierarchical HMM Implementation for Vertebrate Gene Splice Site Prediction. Technical report, Dept. Computer Science, Univ. Waterloo, 2000. [Hoe01] J. Hoey. Hierarchical unsupervised learning of facial expression categories. In ICCV Workshop on Detection and Recognition of Events in Video, 2001. [IB00] Y. Ivanov and A. Bobick. Recognition of visual activities and interactions by stochastic parsing. IEEE Trans. on Pattern Analysis and Machine Intelligence, 22(8):852?872, 2000. [JGS96] M. I. Jordan, Z. Ghahramani, and L. K. Saul. Hidden Markov decision trees. In NIPS, 1996. [Kja90] U. Kjaerulff. Triangulation of graphs ? algorithms giving small total state space. Technical Report R-90-09, Dept. of Math. and Comp. Sci., Aalborg Univ., Denmark, 1990. [LY90] K. Lari and S. J. Young. The estimation of stochastic context-free grammars using the Inside-Outside algorithm. Computer Speech and Language, 4:35?56, 1990. [ME01] D. Moore and I. Essa. Recognizing multitasked activities using stochastic context-free grammar. In CVPR Workshop on Models vs Exemplars in Computer Vision, 2001. [Mur99] K. Murphy. Pearl?s algorithm and multiplexer nodes. Technical report, U.C. Berkeley, Dept. Comp. Sci., 1999. [Mur01] K. Murphy. Applying the junction tree algorithm to variable-length DBNs. Technical report, Comp. Sci. Div., UC Berkeley, 2001. [MW01] K. Murphy and Y. Weiss. The Factored Frontier Algorithm for Approximate Inference in DBNs. In UAI, 2001. [NI00] A. Nefian and M. Hayes III. Maximum likelihood training of the embedded HMM for face detection and recognition. In IEEE Intl. Conf. on Image Processing, 2000. [PR97] R. Parr and S. Russell. Reinforcement learning with hierarchies of machines. In NIPS, 1997. [Rab89] L. R. Rabiner. A tutorial on Hidden Markov Models and selected applications in speech recognition. Proc. of the IEEE, 77(2):257?286, 1989. [SJ99] L. Saul and M. Jordan. Mixed memory markov models: Decomposing complex stochastic processes as mixture of simpler ones. Machine Learning, 37(1):75?87, 1999. [SPS99] R.S. Sutton, D. Precup, and S. Singh. Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning. Artificial Intelligence, 112:181?211, 1999. [TRM01] G. Theocharous, K. Rohanimanesh, and S. Mahadevan. Learning Hierarchical Partially Observed Markov Decision Process Models for Robot Navigation. In ICRA, 2001. [Zwe97] G. Zweig. Speech Recognition with Dynamic Bayesian Networks. PhD thesis, U.C. Berkeley, Dept. Comp. Sci., 1997.
2050 |@word middle:2 version:5 glue:1 hu:1 multitasked:1 thereby:4 solid:1 recursively:1 initial:1 configuration:1 contains:1 current:1 yet:1 must:5 parsing:1 cpds:5 enables:1 designed:1 v:3 alone:1 generative:1 leaf:1 fewer:1 greedy:1 intelligence:2 selected:1 math:1 node:30 simpler:3 height:1 blackwellized:1 become:1 inside:5 expected:1 themselves:1 terminal:4 ivanov:1 param:1 vertebrate:1 bounded:1 what:1 kind:2 string:4 impractical:1 temporal:3 berkeley:5 every:2 exactly:1 returning:1 murphyk:1 control:4 continually:1 before:1 local:1 theocharous:1 sutton:1 path:1 merge:1 plus:1 specifying:1 shaded:1 hmms:9 limited:1 range:1 practice:1 differs:1 area:1 flatten:1 regular:3 get:4 cannot:1 operator:1 scheduling:1 context:9 applying:3 writing:1 restriction:1 equivalent:3 straightforward:1 starting:1 duration:1 automaton:6 simplicity:2 factored:2 rule:3 nesting:1 his:1 handle:1 dbns:6 hierarchy:8 imagine:1 exact:1 losing:1 trick:1 recognition:7 rec:1 bottom:4 observed:2 enters:4 ensures:1 connected:1 counter:1 rescaled:1 russell:1 ham:1 environment:1 complexity:6 dynamic:4 terminating:1 singh:1 rewrite:1 upon:1 exit:1 expires:1 easily:1 represented:4 alphabet:2 train:2 univ:2 describe:2 effective:1 artificial:1 kevin:1 outside:5 encoded:2 larger:1 valued:1 cvpr:1 say:1 drawing:2 otherwise:3 consume:1 grammar:5 ability:1 statistic:1 final:1 online:3 higherlevel:1 sequence:10 advantage:1 essa:1 clamp:1 interconnected:1 product:2 maximal:1 adaptation:1 interaction:1 bobick:1 entered:1 exploiting:1 parent:8 intl:2 derive:2 illustrate:1 exemplar:1 nonterminal:2 implemented:1 c:1 closely:1 stochastic:10 enable:2 memorized:1 generalization:1 ly90:2 extension:1 frontier:2 hold:1 considered:1 viterbi:1 parr:1 adopt:1 omitted:1 estimation:1 proc:1 currently:1 waterloo:1 him:2 create:1 clearly:1 always:2 gaussian:1 rather:1 lifted:1 surveillance:1 quence:1 likelihood:1 inference:23 abstraction:1 typically:1 eliminate:1 hidden:13 relation:1 koller:1 going:3 subroutine:1 interested:1 semantics:1 upward:2 overall:3 pixel:1 plan:1 spatial:2 special:3 smoothing:1 softmax:1 ference:1 uc:1 once:2 never:2 identical:1 represents:4 look:1 unsupervised:1 patten:1 np:1 cpd:1 report:4 aalborg:1 murphy:4 recogntion:1 friedman:1 freedom:1 detection:2 message:2 possibility:1 navigation:2 mixture:5 yielding:1 chain:5 emit:5 facial:1 tree:10 hmdt:1 re:1 rao:1 ence:1 loopy:1 cost:3 entry:1 recognizing:1 too:1 tishby:1 pal:1 off:2 together:1 precup:1 concrete:1 again:1 aaai:1 thesis:1 dr:1 stochastically:1 creating:1 conf:1 return:2 multiplexer:1 converted:1 satisfy:1 depends:3 root:5 observer:1 doing:1 analyze:1 observing:1 start:3 option:1 complicated:2 shai:1 substructure:1 contribution:1 formed:1 efficiently:2 rabiner:1 generalize:1 bayesian:5 eters:1 comp:4 definition:2 handwriting:1 con:1 emits:2 duplicated:1 segmentation:1 actually:1 higher:2 specify:1 wei:1 furthermore:2 just:2 until:2 sketch:1 horizontal:6 parse:2 christopher:1 expressive:1 replacing:1 unsegmented:1 propagation:2 lack:1 effect:1 hence:5 moore:1 illustrated:1 naftali:1 generalized:1 complete:2 interface:3 image:1 multinomial:2 empirically:2 conditioning:2 discussed:2 marginals:2 interpret:1 enter:3 ai:1 dbn:18 similarly:1 particle:1 language:2 robot:1 add:1 triangulation:1 termed:1 certain:1 binary:1 fortunately:1 somewhat:1 converting:1 signal:3 semi:1 full:4 segmented:1 technical:4 match:1 faster:1 long:1 zweig:1 hhmm:24 prediction:1 controller:1 vision:1 represent:14 addition:3 separately:2 fine:1 diagram:5 unlike:1 bringing:1 jordan:3 emitted:2 call:4 integer:1 backwards:2 intermediate:2 iii:1 easy:1 mahadevan:1 independence:1 topology:2 reduce:1 whether:2 expression:4 reuse:1 returned:2 speech:3 cause:2 action:3 matlab:2 remark:1 boutilier:1 generally:1 se:1 factorial:1 category:1 dna:1 simplest:1 specifies:2 tutorial:1 dotted:1 dummy:1 mega:1 discrete:2 write:3 shall:1 key:1 drawn:1 graph:1 sum:1 run:1 letter:1 powerful:1 family:1 reader:1 parsimonious:2 decision:5 layer:2 bound:1 followed:1 fan:1 activity:2 ahead:1 scfgs:2 flat:1 encodes:1 calling:2 generates:4 speed:1 performing:2 speedup:1 department:1 remain:2 smaller:2 em:4 describes:1 making:1 wherever:1 intuitively:1 iccv:1 hoey:1 taken:1 legal:1 equation:3 lari:1 discus:4 turn:4 count:1 needed:4 know:1 singer:1 tractable:1 end:17 junction:3 gaussians:1 decomposing:1 multiplied:1 apply:1 hierarchical:14 enforce:1 distinguished:1 scfg:1 swamy:1 original:5 top:5 remaining:1 ensure:1 running:2 marginalized:1 exploit:1 giving:2 yoram:1 ghahramani:2 icra:1 sweep:1 occurs:1 usual:1 div:1 thank:1 sci:4 hmm:19 me:1 denmark:1 assuming:2 length:10 remind:1 minimizing:1 optionally:2 equivalently:1 unfortunately:4 rise:1 implementation:2 proper:1 policy:4 perform:1 allowing:1 upper:1 vertical:4 observation:11 markov:15 arc:5 immediate:1 stack:4 venkatesh:2 pair:1 required:2 connection:2 california:1 learned:2 pearl:1 nip:2 trans:1 able:2 below:3 pattern:2 max:1 memory:1 video:1 belief:2 event:1 natural:1 indicator:1 representing:5 xml:1 brief:1 mdps:2 finished:7 text:1 prior:2 acknowledgement:1 embedded:2 fully:1 mixed:1 hhmms:14 filtering:3 proportional:1 kjaerulff:1 degree:1 agent:2 sufficient:1 storing:1 production:8 row:1 last:1 free:5 offline:1 wide:1 saul:2 face:1 midpoint:1 sparse:2 ingram:1 slice:7 depth:4 transition:25 world:2 ending:1 computes:2 forward:2 author:1 reinforcement:2 approximate:4 bui:2 gene:1 clique:5 global:1 incoming:2 uai:2 hayes:1 assumed:1 ahmm:2 subsequence:1 continuous:1 why:1 table:2 learn:3 reasonably:1 rohanimanesh:1 ca:1 complex:1 domain:2 ifif:2 did:1 main:1 whole:2 child:1 allowed:1 site:1 west:2 ff:1 cubic:2 sub:12 position:1 inferring:1 exponential:1 governed:1 crude:1 jtree:5 young:1 splice:1 down:1 specific:1 paskin:2 symbol:9 workshop:2 phd:1 downward:1 conditioned:1 easier:1 simply:1 likely:1 visual:2 tracking:1 partially:1 conditional:1 exposition:1 shared:2 content:1 change:4 hard:2 determined:2 except:2 called:3 total:2 formally:3 internal:1 mark:1 goldszmidt:1 arises:2 dept:4
1,153
2,051
A General Greedy Approximation Algorithm with Applications Tong Zhang IBM T.J. Watson Research Center Yorktown Heights, NY 10598 tzhang@watson.ibm.com Abstract Greedy approximation algorithms have been frequently used to obtain sparse solutions to learning problems. In this paper, we present a general greedy algorithm for solving a class of convex optimization problems. We derive a bound on the rate of approximation for this algorithm, and show that our algorithm includes a number of earlier studies as special cases. 1 Introduction The goal of machine learning is to obtain a certain input/output functional relationship from a set of training examples. In order to do so, we need to start with a model of the functional relationship. In practice, it is often desirable to find the simplest model that can explain the data. This is because simple models are often easier to understand and can have significant computational advantages over more complicated models. In addition, the philosophy of Occam?s Razor implies that the simplest solution is likely to be the best solution among all possible solutions, In this paper, we are interested in composite models that can be expressed as linear combinations of basic models. In this framework, it is natural to measure the simplicity of a composite model by the number of its basic model components. Since a composite model in our framework corresponds to a linear weight over the basic model space, therefore our measurement of model simplicity corresponds to the sparsity of the linear weight representation. In this paper, we are interested in achieving sparsity through a greedy optimization algorithm which we propose in the next section. This algorithm is closely related to a number of previous works. The basic idea was originated in [5], where Jones observed that if a target vector in a Hilbert space is a convex combination of a library of basic vectors, then using with basic library vecgreedy approximation, one can achieve an error rate of tors. The idea has been refined in [1] to analyze the approximation property of sigmoidal functions including neural networks.   The above methods can be regarded as greedy sparse algorithms for functional approximation, which is the noise-free case of regression problems. A similar greedy algorithm can also be used to solve general regression problems under noisy conditions [6]. In addition to regression, greedy approximation can also be applied to classification problems. The resulting algorithm is closely related to boosting [2] under the additive model point of view [3]. This paper shows how to generalize the method in [5, 1] for analyzing greedy algorithms (in their case, for functional approximation problems) and apply it to boosting. Detailed analysis will be given in Section 4. Our method can also be used to obtain sparse kernel representations for regression problems. Such a sparse representation is what support vector regression machines try to achieve. In this regard, the method given in this paper complements some recently proposed greedy kernel methods for Gaussian processes such as [9, 10]. The proposed greedy approximation method can also be applied to other prediction problems with different loss functions. For example, in density estimation, the goal is to find a model that has the smallest negative log-likelihood. A greedy algorithm was analyzed in [7]. Similar approximation bounds can be directly obtained under the general framework proposed in this paper. We proceed as follows. Section 2 formalizes the general class of problems considered in this paper, and proposes a greedy algorithm to solve the formulation. The convergence rate of the algorithm is investigated in Section 3. Section 4 includes a few examples that can be obtained from our algorithm. Some final concluding remarks are given in Section 5. 2 General Algorithm        In machine learning, our goal is often to predict an unobserved output value based on an observed input vector . This requires us to estimate a functional relationship from a set of example pairs of . Usually the quality of the predictor can be measured by a loss function that is problem dependent.       In this paper, we are interested in the following scenario: given a family of basic predictors     parameterized by , we want to obtain a good predictor   that lies in the convex   with the fewest possible terms:            , where   are nonhull of    . This family of models can be regarded as additive negative weights so that     can be regarded as a vector models in statistics [4]. Formally, each basic model  in a linear functional space. Our problem in its most general form can thus be described  to minimize a functional  of  that as to find a vector   in the convex hull of  measures the quality of  . This functional  of  plays the role of loss function for learning problems.  More formally, we consider a linear vector space  , and a subset  ! . Denote by "$#  the convex hull of  : "$#   % '&)+* (-, +/.0+21 , +2354  +* (6, +   .7+ 89:;=<?>A@B where we use < > to denote the set of positive integers.  We consider the following optimization problem on "$#  : CDFE  Q $R (1) GIHKJML$NOBP   In this paper, we assume that  is a differentiable convex function on "/#  . We propose the following algorithm to approximately solved (1). Algorithm 2.1 (Sparse greedy approximation) Q    8 "/ # R R/ R Q  = and 4    that minimize Q      ,   Q  M  Q    ,  ,  Q   Q      let  given  for find   , end ,    For simplicity, we assume that the minimization of in Algorithm 2.1 can be exactly achieved at each step. This assumption is not essential, and can be easily removed using a slightly more refined analysis. However due to the space limitation, we shall not consider this generalization.   Q %   Q G HKCD7JML/E NOBP   Q $R For convenience, we introduce the following quantity      Q  4   In next as  section, we show that under appropriate regularity conditions,  the  , where is computed from Algorithm 2.1. In addition, the convergence rate can be bounded as . Q      3 Approximation bound   Given any convex function , we have the following proposition, which is a direct consequence of the definition of convexity. In convex analysis, The gradient  can be replaced by the concept of subgradient, which we do not consider in this paper for simplicity.  Proposition 3.1 Consider a convex function , and two vectors and , we have       where    Q   Q   Q 3  Q Q   Q   is the gradient of  . Q Q  The following lemma is the main theoretical result of the paper, which bounds the performance of each greedy solution step in Algorithm 2.1. We assume that is second order differentiable.  Lemma 3.1 Let  "! Q G G KH JML$NOBP    Q 3 MQ CDFE H  G HO    Q  Q  $ #    $ of exists everywhere in where we   that  :assume &the% Hessian if , we have     ,  + +    ' )( # * #     &% if , we have     ., + +      / ' )( # * #  Q !"$#   Q H CDFE G HO  Q 8"$#   Q !  4    ,+ MQ +BQ    Q MQ Q Q "$#   . For all vectors  Q $ R  Q  Proof. Using Taylor expansion+ and the21 definition of for all , , and 0 , Q  +  3 , we have the following inequality Q     Q + $   %  $R Q + = Q+ , * ( +    ,+ Q + Q +      Q  +  * ( + , + , +2354Q It is easy to see that this implies the inequality CDF+ E    + MQ + Q +     Q  +  +(  +  , R   R R/R  :  Now, consider two sequences and (   Multiply the above inequality (with replaced by ) by + ), such that . , and sum over , we obtain   +    $  , + Q + Q )  Q A * ( + Q +   Q     Q + ,  + $  R Using Proposition 3.1, we obtain  CDF+ E    + MQ + Q+     Q  +   * ( + Q+     Q  + $ R + , Q  Q Since in the above, + and + are arbitrary, therefore +(  + + can be used to express Q  , , any vector 8"$#  . This implies  CDFE    ,+ MQ + Q     Q +  CDF E   Q     Q  + $ R G H O G +        Q  CDFE   Q   % in the above inequality, we obtain  C D G  Now by setting    the lemma.    Q    Using the above lemma and note that  , it is easy to obtain the following theorem by induction. For space limitation, we skip the proof.  3    Q  Theorem 3.1 Under the assumptions of Lemma 3.1, Algorithm 2.1 approximately solves (1), and the rate of convergence for is given by /      &%  If , then we also have /     Q   R    Q     N G P R 4 Examples In this section, we discuss the application of Algorithm 2.1 in some learning problems. We show that the general formulation considered in this paper includes some previous formulations as special cases. We will also compare our results with similar results in the literature. 4.1 Regression    so that the expected loss of       #       $ In regression, we would like to approximate as    is small, where we use the squared loss for simplicity (this choice is obviously not crucial # is the expectation over and , which often corresponds to the in our framework). empirical distribution of pairs. It may also represent the true distribution for some       other engineering applications. Given a set of basis functions with , we may consider the following regression formulation that is slightly different from (1):  CDF E   #    *         $  *        s.t.    (2) where is a positive regularization parameter which is used to control the size of the weight vector . The above formulation can be readily converted into (1) by considering the following set of basic vectors:     &     1       @ R 4 (Q 4 ) in Algorithm 2.1. Since the quantity in Lemma 3.1 We may start with  can be bounded as   "!   $     $ R Q   This implies that the sparse solution in Algorithm 2.1, represented as weight    / R R R  and (   ), satisfies the following inequality:    #    *       $  CDFE     #    * (  +   +  $   $  3!       $  #  +   for all 3 . This leads to the original functional approximation results in [1, 5] and its generalization in [6]. &    /R R R  6@ The sparse regression algorithm studied in this section can also be applied to kernel methods. In this case, corresponds to the input training data space , and the basis . Clearly, this corresponds to a special case of predictors are of the form (2). A sparse kernel representation can be obtained easily from Algorithm 2.1 which leads to provably good approximation rate. Our sparse kernel regression formulation is related to Gaussian processes, where greedy style algorithms have also been proposed [9, 10]. The bound given here is comparable to the bound given in [10] where a sparse approximation rate of the form was obtained.             4.2 Binary classification and Boosting    &   @ In binary classification, the output value is a discrete variable. Given a continuous model , we consider the following prediction rule:    3 4 R    5   4  4 , which is assumed to occur The classification error (we shall ignore the point   rarely) can be given by         if    54 R 4 if    4 if if Unfortunately, this classification error function is not convex, which cannot be handled in our formulation. In fact, even in many other popular methods, such as logistic regression and support vector machines, some kind of convex formulations have to be employed. Although it is possible for us to analyze their formulations, in this section, we only consider the following form of loss that is closely related to Adaboost [2]: !  # (3)       D            where  is a scaling factor.    0    21 , which are often called weak Again, we consider a set of basis predictors   learners in the boosting literature. We would like to find a strong learner   as a convex combination of weak learners to approximately minimize the above loss: CD7E D    #   !   *        (4)    *         354 R s.t. (5)   &      1 4   This can be written as formulation (1) with  @R                $   R $   HKJML$NOBP           4 in Algorithm 2.1. Theorem 3.1 implies that the sparse solution Q  , We start with     R R/R   ), satisfies the following inequality: represented as weight  and  (   #   !    *         C DF E     #   !   * (  +   +   /  $   #  + Using simple algebra, it is easy to verify that  !  #  "! $  !  # #   . for all 3  (6)   4 Weight in the above inequality is non-negative. Now we consider the special situation that there exists such that  + 354  + *  C  DF E  #    #   !    + (  +   +      !      $R   4 (7)       * (  +   +   3   R + This condition will be satisfied in the large margin linearly separable case where there exists   and such that and for all data , Now, under (7), we obtain from (6) that  *             !     /  $  $R          to obtain Fix any 3 , we can choose   *             !   $       R   (8) This implies that the misclassification error rate decays exponentially. The exponential decay of misclassification error is the original motivation of Adaboost [2]. Boosting was later viewed as greedy approximation in the additive model framework [3]. From the learning theory perspective, the good generalization ability of boosting is related to its tendency to improve the misclassification error under a positive margin [8]. From this point of view, inequality (8) gives a much more explicit margin error bound (which decreases exponentially) than a related result in [8]. In the framework of additive models, Adaboost corresponds to the exponential loss (3) analyzed in this section. As pointed out in [3], other loss functions can also be used. Using our analysis, we may also obtain sparse approximation bounds for these different loss functions. However, it is also easy to observe that they will not lead to the exponential decay of classification error in the separable case. Although the exponential loss in (3) is attractive for separable problems due to the exponential decay of margin error, it is very sensitive to outliers in the non-separable case. We shall mention that an interesting aspect of boosting is the concept of adaptive resampling or sample reweighting. Although this idea has dominated the interpretation of boosting algorithms, it has been argued in [3] that adaptive resampling is only a computational by-product. The idea corresponds to a Newton step approximation in the sparse greedy solution of in Algorithm 2.1 under the additive model framework which we consider here. Our analysis further confirmed that the greedy sparse solution of an additive model in (1), rather than reweighting itself is the key component in boosting. In our framework, it is also much easier to related the idea of boosting to the greedy function approximation method outlined in [1, 5].   4.3 Mixture density estimation  In mixture density estimation, the output is the probability density function of the input vector at . The following negative log-likelihood is commonly used as loss function:         D      where   3 4 is a probability density function.   , which are often called mixture comAgain, we consider a set of basis predictors   ponents. We would like to find a mixture probability density model   as a convex combination of mixture components to approximately minimize the negative log-likelihood: CDFE    D  *        (9)     *    %3 4 R  s.t. (10)   This problem was studied in [7]. The quantity defined in Lemma 3.1 can be computed as:   "!   "! $ $ #   #    $ $ $ $ An approximation bound can now be directly obtained from Theorem 3.1. It has a form similar to the bound given in [7].             R   N P N P HKJML/NOBP      5 Conclusion This paper studies a formalization of a general class of prediction problems in machine learning, where the goal is to approximate the best model as a convex combination of a family of basic models. The quality of the approximation can be measured by a loss function which we want to minimize.     We proposed a greedy algorithm to solve the problem, and we have shown that for a variety of loss functions, a convergence rate of can be achieved using a convex combination of basic models. We have illustrated the consequence of this general algorithm in regression, classification and density estimation, and related the resulting algorithms to previous methods.  References [1] A.R. Barron. Universal approximation bounds for superpositions of a sigmoidal function. IEEE Transactions on Information Theory, 39(3):930?945, 1993. [2] Y. Freund and R.E. Schapire. A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci., 55(1):119?139, 1997. [3] Jerome Friedman, Trevor Hastie, and Robert Tibshirani. Additive logistic regression: A statistical view of boosting. The Annals of Statistics, 28(2):337?407, 2000. With discussion. [4] T. J. Hastie and R. J. Tibshirani. Generalized additive models. Chapman and Hall Ltd., London, 1990. [5] Lee K. Jones. A simple lemma on greedy approximation in Hilbert space and convergence rates for projection pursuit regression and neural network training. Ann. Statist., 20(1):608?613, 1992. [6] Wee Sun Lee, P.L. Bartlett, and R.C. Williamson. Efficient agnostic learning of neural networks with bounded fan-in. IEEE Transactions on Information Theory, 42(6):2118?2132, 1996. [7] Jonathan Q. Li and Andrew R. Barron. Mixture density estimation. In S.A. Solla, T.K. Leen, and K.-R. M?uller, editors, Advances in Neural Information Processing Systems 12, pages 279?285. MIT Press, 2000. [8] Robert E. Schapire, Yoav Freund, Peter Bartlett, and Wee Sun Lee. Boosting the margin: a new explanation for the effectiveness of voting methods. Ann. Statist., 26(5):1651?1686, 1998. [9] Alex J. Smola and Peter Bartlett. Sparse greedy Gaussian process regression. In Advances in Neural Information Processing Systems 13, pages 619?625, 2001. [10] Tong Zhang. Some sparse approximation bounds for regression problems. In The Eighteenth International Conference on Machine Learning, pages 624?631, 2001.
2051 |@word concept:2 skip:1 implies:6 true:1 verify:1 regularization:1 strong:1 closely:3 quantity:3 hull:2 illustrated:1 attractive:1 gradient:2 mention:1 razor:1 argued:1 yorktown:1 sci:1 originated:1 fix:1 generalization:4 generalized:1 proposition:3 theoretic:1 induction:1 com:1 considered:2 hall:1 relationship:3 obp:5 recently:1 written:1 readily:1 predict:1 unfortunately:1 additive:8 functional:9 robert:2 tor:1 smallest:1 negative:5 exponentially:2 estimation:5 resampling:2 interpretation:1 greedy:22 superposition:1 significant:1 measurement:1 sensitive:1 outlined:1 minimization:1 uller:1 mit:1 boosting:13 clearly:1 gaussian:3 pointed:1 situation:1 sigmoidal:2 rather:1 zhang:2 arbitrary:1 height:1 direct:1 complement:1 pair:2 perspective:1 likelihood:3 introduce:1 scenario:1 certain:1 inequality:8 expected:1 binary:2 watson:2 frequently:1 dependent:1 usually:1 employed:1 sparsity:2 considering:1 interested:3 including:1 bounded:3 provably:1 desirable:1 agnostic:1 among:1 classification:7 what:1 explanation:1 kind:1 proposes:1 misclassification:3 natural:1 special:4 hkc:1 unobserved:1 improve:1 formalizes:1 chapman:1 library:2 prediction:3 jones:2 voting:1 basic:11 regression:16 expectation:1 exactly:1 df:7 kernel:5 control:1 few:1 represent:1 achieved:2 literature:2 wee:2 addition:3 positive:3 want:2 engineering:1 freund:2 loss:14 replaced:2 consequence:2 jml:1 crucial:1 interesting:1 limitation:2 analyzing:1 friedman:1 approximately:4 multiply:1 studied:2 effectiveness:1 integer:1 analyzed:2 mixture:6 editor:1 explicit:1 occam:1 ibm:2 easy:4 variety:1 free:1 hastie:2 practice:1 idea:5 understand:1 bq:1 taylor:1 sparse:16 universal:1 empirical:1 handled:1 bartlett:3 composite:3 projection:1 theoretical:1 ltd:1 regard:1 earlier:1 peter:2 commonly:1 adaptive:2 convenience:1 cannot:1 proceed:1 yoav:1 hessian:1 remark:1 transaction:2 approximate:2 subset:1 detailed:1 ignore:1 predictor:6 center:1 eighteenth:1 statist:2 convex:15 simplest:2 assumed:1 simplicity:5 schapire:2 rule:1 continuous:1 density:8 regarded:3 international:1 tibshirani:2 mq:6 lee:3 discrete:1 shall:3 annals:1 target:1 play:1 express:1 squared:1 again:1 satisfied:1 key:1 achieving:1 choose:1 expansion:1 williamson:1 investigated:1 main:1 linearly:1 motivation:1 noise:1 style:1 subgradient:1 li:1 syst:1 sum:1 converted:1 observed:2 role:1 parameterized:1 everywhere:1 solved:1 includes:3 tzhang:1 family:3 ny:1 sun:2 tong:2 formalization:1 solla:1 decrease:1 removed:1 later:1 view:3 try:1 decision:1 analyze:2 scaling:1 convexity:1 start:3 comparable:1 bound:12 complicated:1 exponential:5 comput:1 lie:1 fan:1 theorem:4 solving:1 minimize:5 algebra:1 occur:1 alex:1 decay:4 learner:3 basis:4 easily:2 generalize:1 weak:2 dominated:1 aspect:1 represented:2 concluding:1 essential:1 exists:3 separable:4 fewest:1 confirmed:1 london:1 margin:5 easier:2 explain:1 combination:6 refined:2 trevor:1 slightly:2 definition:2 likely:1 solve:3 expressed:1 ability:1 statistic:2 proof:2 outlier:1 noisy:1 itself:1 final:1 corresponds:7 obviously:1 popular:1 advantage:1 differentiable:2 sequence:1 satisfies:2 discus:1 goal:4 propose:2 hilbert:2 viewed:1 product:1 ann:2 ponents:1 end:1 philosophy:1 pursuit:1 adaboost:3 achieve:2 apply:1 observe:1 formulation:10 leen:1 barron:2 kh:1 appropriate:1 lemma:8 called:2 smola:1 tendency:1 jerome:1 convergence:5 regularity:1 ho:2 rarely:1 original:2 formally:2 support:2 reweighting:2 jonathan:1 derive:1 andrew:1 logistic:2 newton:1 quality:3 measured:2 dfe:6 solves:1
1,154
2,052
Variance Reduction Techniques for Gradient Estimates in Reinforcement Learning Evan Greensmith Australian National University evan@csl.anu.edu.au Peter L. Bartlett? BIOwulf Technologies Peter.Bartlett@anu.edu.au Jonathan Baxter? WhizBang! Labs, East jbaxter@whizbang.com Abstract We consider the use of two additive control variate methods to reduce the variance of performance gradient estimates in reinforcement learning problems. The first approach we consider is the baseline method, in which a function of the current state is added to the discounted value estimate. We relate the performance of these methods, which use sample paths, to the variance of estimates based on iid data. We derive the baseline function that minimizes this variance, and we show that the variance for any baseline is the sum of the optimal variance and a weighted squared distance to the optimal baseline. We show that the widely used average discounted value baseline (where the reward is replaced by the difference between the reward and its expectation) is suboptimal. The second approach we consider is the actor-critic method, which uses an approximate value function. We give bounds on the expected squared error of its estimates. We show that minimizing distance to the true value function is suboptimal in general; we provide an example for which the true value function gives an estimate with positive variance, but the optimal value function gives an unbiased estimate with zero variance. Our bounds suggest algorithms to estimate the gradient of the performance of parameterized baseline or value functions. We present preliminary experiments that illustrate the performance improvements on a simple control problem. 1 Introduction, Background, and Preliminary Results In reinforcement learning problems, the aim is to select a controller that will maximize the average reward in some environment. We model the environment as a partially observable Markov decision process (POMDP). Gradient ascent methods (e.g., [7, 12, 15]) estimate the gradient of the average reward, usually using Monte Carlo techniques to cal? Most of this work was performed while the authors were with the Research School of Information Sciences and Engineering at the Australian National University. culate an average over a sample path of the controlled POMDP. However such estimates tend to have a high variance, which means many steps are needed to obtain a good estimate. GPOMDP [4] is an algorithm for generating an estimate of the gradient in this way. Compared with other approaches, it is suitable for large systems, when the time between visits to a state is large but the mixing time of the controlled POMDP is short. However, it can suffer from the problem of producing high variance estimates. In this paper, we investigate techniques for variance reduction in GPOMDP. One generic approach to reducing the variance of Monte Carlo estimates of integrals is to use an additive control variate (see, for example, [6]). Suppose we wish to estimate the integral ofR f : X ? R, Rand we know the R integral of another function ? : X ? R. Since X f = X (f ? ?) + X ?, the integral of f ? ? can be estimated instead. Obviously if ? = f then the variance is zero. More generally, Var(f ? ?) = Var(f ) ? 2Cov(f, ?) + Var(?), so that if ? and f are strongly correlated, the variance of the estimate is reduced. In this paper, we consider two approaches of this form. The first (Section 2) is the technique of adding a baseline. We find the optimal baseline and we show that the additional variance of a suboptimal baseline can be expressed as a weighted squared distance from the optimal baseline. Constant baselines, which do not depend on the state or observations, have been widely used [13, 15, 9, 11]. In particular, the expectation over all states of the discounted value of the state is a popular constant baseline (where, for example, the reward at each step is replaced by the difference between the reward and the expected reward). We give bounds on the estimation variance that show that, perhaps surprisingly, this may not be the best choice. The second approach (Section 3) is the use of an approximate value function. Such actorcritic methods have been investigated extensively [3, 1, 14, 10]. Generally the idea is to minimize some notion of distance between the fixed value function and the true value function. In this paper we show that this may not be the best approach: selecting the fixed value function to be equal to the true value function is not always the best choice. Even more surprisingly, we give an example for which the use of a fixed value function that is different from the true value function reduces the variance to zero, for no increase in bias. We give a bound on the expected squared error (that is, including the estimation variance) of the gradient estimate produced with a fixed value function. Our results suggest new algorithms to learn the optimum baseline, and to learn a fixed value function that minimizes the bound on the error of the estimate. In Section 5, we describe the results of preliminary experiments, which show that these algorithms give performance improvements. POMDP with Reactive, Parameterized Policy A partially observable Markov decision process (POMDP) consists of a state space, S, a control space, U, an observation space, Y, a set of transition probability matrices {P(u) : u ? U}, each with components pij (u) for i, j ? S, u ? U, an observation process ? : S ? PY , where PY is the space of probability distributions over Y, and a reward function r : S ? R. We assume that S, U, Y are finite, although all our results extend easily to infinite U and Y, and with more restrictive assumptions can be extended to infinite S. A reactive, parameterized policy for a POMDP is a set of mappings {?(?, ?) : Y ? PU |? ? RK }. Together with the POMDP, this defines the controlled POMDP (S, U, Y, P , ?, r, ?). The joint state, observation and control process, {Xt , Yt , Ut }, is Markov. The state process, {Xt }, is also Markov, with transition probP abilities pij (?) = y?Y,u?U ?y (i)?u (y, ?)pij (u), where ?y (i) denotes the probability of observation y given the state i, and ?u (y, ?) denotes the probability of action u given parameters ? and observation y. The Markov chain M(?) = (S, P(?)) then describes the behaviour of the process {Xt }. Assumption 1 The controlled POMDP (S, U, Y, P , ?, r, ?) satisfies: For all ? ? RK there exists a unique stationary distribution satisfying ? 0 (?) P(?) = ? 0 (?). There is an R < ? such that, for all i ? S, |r(i)| ? R. There is a B < ? such that, for all u ? U, y ? Y and ? ? RK the derivatives ??u (y, ?)/??k (1 ? k ? K) exist, and the vector of these derivatives satisfies k??u (y, ?)/?u (y, ?)k ? B, where k ? k denotes the Euclidean norm on RK . h P i def T ?1 We consider the average reward, ?(?) = limT ?? E T1 t=0 r(Xt ) . Assumption 1 implies that this limit exists, and does not depend on the start state X0 . The aim is to def select a policy quantity.i Define the discounted value function, J ? (i, ?) = h Pto maximize this T ?1 t limT ?? E t=0 ? r(Xt ) X0 = i . Throughout the rest of the paper, dependences upon ? are assumed, and dropped in the notation. For a random vector A, we denote h i Var(A) = E (A ? E [A])2 , where a2 denotes a0 a, and a0 denotes the transpose of the column vector a. GPOMDP Algorithm The GPOMDP h algorithm i[4] uses a sample path to estimate the gradient approximation def u(y) ?? ? = E ?? ?u(y) J? (j) . As ? ? 1, ?? ? approaches the true gradient ??, but the def P2T variance increases. We consider a slight modification [2]: with Jt = s=t ? s?t r(Xs ), def ?T = T ?1 1 X ??Ut (Yt ) Jt+1 . T t=0 ?Ut (Yt ) (1) Throughout this paper the process {Xt , Yt , Ut , Xt+1 } is generally understood to be generated by a controlled POMDP satisfying Assumption 1, with X0 ?? (ie the initial state distributed according to the stationary distribution). That is, before computing the gradient estimates, we wait until the process has settled down to the stationary distribution. Dependent Samples Correlation terms arise in the variance quantities to be analysed. We show here that considering iid samples gives an upper bound on the variance of the general case. The mixing time of a finite ergodic Markov chain M = (S, P ) is defined as        def ? = min t > 1 : max dT V P t i , P t j ? e?1 , i,j t t where [P ]i denotes the ith row of P and dT V is the total variation distance, dT V (P, Q) = P |P (i) ? Q(i)|. i Theorem 1 Let M = (S, P ) be a finite ergodic Markov chain, p with mixing time ? , and 2|S|e and 0 ? ? < let ? be its stationary distribution. There are constants L < exp(?1/(2? )), which depend only on M , such that, for all f : S ? R and all t, Cov?f (t) ? L?t Var? (f), where Var? (f) is the variance of f under ?, and Cov?f (t) is the auto-covariance of the process {f(Xt )}, where the process {Xt } is generated by M with initial distribution ?. Hence, for some constant ?? ? 4L? , ! T ?1 1 X ?? Var Var? (f). f(Xt ) ? T t=0 T We call (L, ? ) the mixing constants of the Markov chain M (or of the controlled POMDP D; ie the Markov chain (S, P )). We omit the proof (all proofs are in the full version [8]). Briefly, we show that for a finite ergodic Markov chain M , Cov?f (t) ? Rt (M )Var? (f),   2 for some Rt (M ). We then show that Rt (M ) < 2|S| exp(? ?t ). In fact, for a reversible chain, we can choose L = 1 and ? = |?2 |, the second largest magnitude eigenvalue of P . 2 Baseline We consider an alteration of (1), def ?T = T ?1 1 X ??Ut (Yt ) (Jt+1 ? Ar (Yt )) . T t=0 ?Ut (Yt ) (2) For any baseline Ar : Y ? R, it is easy to show that E [?T ] = E [?T ]. Thus, we select Ar to minimize variance. The following theorem shows that this variance is bounded by a variance involving iid samples, with Jt replaced by the exact value function. Theorem 2 Suppose that D = (S, U, Y, P , ?, r, ?) is a controlled POMDP satisfying Assumption 1, D has mixing constants (L, ? ), {Xt , Yt , Ut , Xt+1 } is a process generated by D with X0 ?? ,Ar P : Y ? R is a baseline that is uniformly bounded by M, and J (j) s has the distribution of ? s=0 ? r(Xt ), where the states Xt are generated by D starting in X0 = j. Then there are constants C ? 5B2 R(R + M) and ? ? 4L? ln(eT ) such that !   T ?1 1 X ??Ut (Yt ) ? ??u (y) Var (Jt+1 ?Ar (Yt )) ? Var? (J? (j)?Ar (y)) T t=0 ?Ut (Yt ) T ?u (y)  2   ? ??u (y) ? C + E (J (j) ? J? (j)) + +1 ?T , T ?u (y) T (1 ? ?)2 where, as always, (i, y, u, j) are generated iid with i??, y??(i), u??(y) and j?P i (u). The proof uses Theorem 1 and [2, Lemma 4.3]. Here we have bounded the variance of (2) with the variance of a quantity we may readily analyse. The second term on the right hand side shows the error associated with replacing an unbiased, uncorrelated estimate of the value function with the true value function. This quantity is not dependent on the baseline. The final term on the right hand side arises from the truncation of the discounted reward? and is exponentially decreasing. We now concentrate on minimizing the variance   ??u (y) 2 (J? (j) ? Ar (y)) , (3) ? r = Var? ?u (y) which the following lemma relates to the variance ? 2 without a baseline,   ??u (y) J? (j) . ? 2 = Var? ?u (y) Lemma 3 Let D = (S, U, Y, P , ?, r, ?) be a controlled POMDP satisfying Assumption 1. For any baseline Ar : Y ? R, and for i??, y??(i), u??(y) and j?Pi (u), ## " " " 2 # 2 ??u (y) ??u (y) 2 2 2 ? r = ? + E Ar (y) E J? (j) y . y ? 2Ar (y)E ?u (y) ?u (y) From Lemma 3 it can be seen that the task of finding the optimal baseline is in effect that of minimizing a quadratic for each observation y ? Y. This gives the following theorem. Theorem 4 For the controlled POMDP as in Lemma 3, ? #!2 " #? " 2 2 ?? (y) ?? (y) u min ? 2r = ? 2 ? E ? E J? (j) y /E ? u(y) y ? , Ar ?u (y) u and this minimum is attained with the baseline # " # " 2 2 ?? (y) ?? (y) u u ? J? (j) y /E Ar (y) = E y . ?u (y) ?u (y) Furthermore, the optimal constant baseline is " # 2  2 ??u (y) ??u (y) ? Ar = E J? (j) /E . ?u (y) ?u (y) The following theorem shows that the variance of an estimate with an arbitrary baseline can be expressed as the sum of the variance with the optimal baseline and a certain squared weighted distance between the baseline function and the optimal baseline function. Theorem 5 If Ar : Y ? R is a baseline function, A?r is the optimal baseline defined in Theorem 4, and ? 2r? is the variance of the corresponding estimate, then " # 2 ?? (y) u 2 ? 2r = ? 2r? + E (Ar (y) ? A?r (y)) , ?u (y) where i??, y ??(i), and u??(y). Furthermore, the same result is true for the case of constant baselines, with Ar (y) replaced by an arbitrary constant baseline Ar , and A?r (y) replaced by A?r , the optimum constant baseline defined in Theorem 4. For the constant baseline Ar = E i?? [J? (i)], Theorem 5 implies that ? 2r is equal to " #!2 2 2  2  ??u (y) ??u (y) ??u (y) 2 min ? r + E E [J? (j)] ? E J? (j) /E ? (y) . Ar ?R ?u (y) ?u (y) u Thus, its performance depends on the random variables (??u (y)/?u (y))2 and J? (j); if they are nearly independent, E [J? ] is a good choice. 3 Fixed Value Functions: Actor-Critic Methods We consider an alteration of (1), T ?1 1 X ??Ut (Yt ) ? ? T def ? = J? (Xt+1 ), T t=0 ?Ut (Yt ) ?? : S ? R. Define for some fixed value function J " ? # X def k A? (j) = E ? d(Xk , Xk+1 ) X0 = j , (4) k=0 def ?? (j) ? J ?? (i) is the temporal difference. Then it is easy to show where d(i, j) = r(i) + ? J that the estimate (4) has a bias of   h i ? T = E ??u (y) A? (j) . ?? ? ? E ? ?u (y) The following theorem gives a bound on the expected squared error of (4). The main tool in the proof is Theorem 1. Theorem 6 Let D = (S, U, Y, P , ?, r, ?) be a controlled POMDP satisfying Assumption 1. For a sample path from D, that is, {X0??, Yt??(Xt ), Ut??(Yt ), Xt+1?PXt (Ut )},     2  2  ? ? ??u (y) ? ??u (y) ? T ? ?? ? Var? J? (j) + E A? (j) , E ? ? T ?u (y) ?u (y) where the second expectation is over i??, y??(i), u??(y), and j?P i (u). ?? (j) = J? (j) + v(j), then by selecting v = (v(1), . . . , v(|S|))0 from the right If we write J def P 0 null space of the K ? |S| matrix G, where G = i,y,u ?i ?y (i)??u (y)Pi (u), (4) will produce an unbiased estimate of ?? ?. An obvious example of such a v is a constant vector, (c, c, . . . , c)0 : c ? R. We can use this to construct a trivial example where (4) produces an unbiased estimate with zero variance. Indeed, let D = (S, U, Y, P , ?, r, ?) be a controlled POMDP satisfying Assumption 1, with r(i) = c, for some 0 < c ? R. Then J? (j) = c/(1 ? ?) and ?? ? = 0. If we choose v = (?c/(1 ? ?), . . . , ?c/(1 ? ?))0 and ?? (j) = J? (j) + v(j), then ??u(y) J ? J ?u(y) ? (j) = 0 for all y, u, j, and so (4) gives an unbiased estimate of ?? ?, with zero variance. Furthermore, for any D for which there exists a pair y, u such that ?u (y) > 0 and ??u (y) 6= 0, choosing ? J? (j) = J? (j) gives a variance greater than zero?there is a non-zero probability event, (Xt = i, Yt = y, Ut = u, Xt+1 = u(y) j), such that ?? ?u(y) J? (j) 6= 0. 4 Algorithms  Given a parameterized class of baseline functions Ar (?, ?) : Y ? R ? ? RL , we can use Theorem 5 to bound the variance of our estimates. Computing the gradient of this bound with respect to the parameters ? of the baseline function allows a gradient optimization of the baseline. The GDORB Algorithm produces an estimate ? S of these gradients from a sample path of length S. Under the assumption that the baseline function and its gradients are uniformly bounded, we can show that these estimates converge to the gradient of ? 2r . We omit the details (see [8]). GDORB Algorithm: Given: Controlled POMDP (S, U, Y, P , ?, r, ?), parameterized baseline Ar . set z0 = 0 (z0 ? RL ), ?0 = 0 (?0 ? RL ) for all {is , ys , us , is+1 , ys+1 } generated by the POMDP do   ?? s(ys ) 2 zs+1 = ?zs + ?Ar (ys ) ?uu(y ) s s ?s+1 = ?s + end for 1 s+1 ((Ar (ys ) ? ?Ar (ys+1 ) ? r(xs+1 )) zs+1 ? ?s ) ?? (?, ?) : S ? R ? ? RL }, we can For a parameterized class of fixed value functions {J use Theorem 6 to bound the expected squared error of our estimates, and compute the gradient of this bound with respect to the parameters ? of the baseline function. The GBTE Algorithm produces an estimate ?S of these gradients from a sample path of length S. Under the assumption that the value function and its gradients are uniformly bounded, we can show that these estimates converge to the true gradient. GBTE Algorithm: Given: Controlled POMDP (S, U, Y, P , ?, r, ?), parameterized fixed value function ? J? . set z0 = 0 (z0 ? RK ), ?A0 = 0 (?A0 ? R1?L ), ?B 0 = 0 (?B 0 ? RK ), ?C 0 = 0 (?C 0 ? RK ) and ?D0 = 0 (?D0 ? RK?L ) for all {is , ys , us , is+1 , is+2 } generated by the POMDP do ?? s(ys ) zs+1 = ?zs + ?uu(y s s)    0   0  ??us(ys ) ? ??us(ys ) 1 ? ?As+1 = ?As + s+1 ?J? (is+1 ) ? ?As ?us(ys ) J? (is+1 ) ?us(ys )   ??us(ys ) ? 1 J? (is+1 ) ? ?B s ?B s+1 = ?B s + s+1 ?us(ys )   1 ?? (is+2 ) ? J ?? (is+1 ) zs+1 ? ?C s r(is+1 ) + ? J ?C s+1 = ?C s + s+1    0 ??us(ys ) 1 ? ?Ds+1 = ?Ds + s+1 ?u (ys ) ?J? (is+1 ) ? ?D s s 0  ? ?? ? ?s+1 = T ?As+1 ? T ?B 0s+1 ?D s+1 ? ?C 0s+1 ?Ds+1 end for 5 Experiments Experimental results comparing these GPOMDP variants for a simple three state MDP (described in [5]) are shown in Figure 1. The exact value function plots show how different choices of baseline and fixed value function compare when all algorithms have access to the exact value function J? . Using the expected value function as a baseline was an improvement over GPOMDP. Using the optimum, or optimum constant, baseline was a further improvement, each performing comparably to the other. Using the pre-trained fixed value function was also an improvement over GPOMDP, showing that selecting the true value function was indeed not the best choice in this case. The trained fixed value function was not optimal though, as J? (j) ? A?r is a valid choice of fixed value function. The optimum baseline, and fixed value function, will not normally be known. The online plots show experiments where the baseline and fixed value function were trained using online gradient descent whilst the performance gradient was being estimated, using the same data. Clear improvement over GPOMDP is seen for the online trained baseline variant. For the online trained fixed value function, improvement is seen until T becomes?given the simplicity of the system?very large. References [1] L. Baird and A. Moore. Gradient descent for general reinforcement learning. In Advances in Neural Information Processing Systems 11, pages 968?974. MIT Press, 1999. [2] P. L. Bartlett and J. Baxter. Estimation and approximation bounds for gradient-based reinforcement learning. Journal of Computer and Systems Sciences, 2002. To appear. [3] A. G. Barto, R. S. Sutton, and C. W. Anderson. Neuronlike adaptive elements that can solve difficult learning control problems. IEEE Transactions on Systems, Man, and Cybernetics, SMC-13:834?846, 1983. [4] J. Baxter and P. L. Bartlett. Infinite-horizon gradient-based policy search. Journal of Artificial Intelligence Research, 15:319?350, 2001. [5] J. Baxter, P. L. Bartlett, and L. Weaver. Infinite-horizon gradient-based policy search: II. Gradient ascent algorithms and experiments. Journal of Artificial Intelligence Research, 15:351?381, 2001. [6] M. Evans and T. Swartz. Approximating integrals via Monte Carlo and deterministic methods. Oxford University Press, 2000. Exact Value Function?Mean Error Exact Value Function?One Standard Deviation 0.4 0.4 GPOMDP-J? BL- [J? ] BL-A?r (y) BL-A?r FVF-pretrain 0.3 0.25 GPOMDP-J? BL- [J? ] BL-A?r (y) BL-A?r FVF-pretrain 0.35 Relative Norm Difference Relative Norm Difference 0.35 0.2 0.15 0.1 0.05  0.3 0.25 0.2 0.15 0.1 0.05 0 0 1 10 100 1000 10000 100000 1e+06 1e+07 1 10 100 1000 T Online?Mean Error 100000 1e+06 1e+07 Online?One Standard Deviation 1 1 GPOMDP BL-online FVF-online 0.8 Relative Norm Difference Relative Norm Difference 10000 T 0.6 0.4 0.2 0 GPOMDP BL-online FVF-online 0.8 0.6 0.4 0.2 0 1 10 100 1000 10000 100000 1e+06 1e+07 1 10 100 T 1000 10000 100000 1e+06 1e+07 T Figure 1: Three state experiments?relative norm error k? est ? ??k / k??k. Exact value function plots compare mean error and standard deviations for gradient estimates (with knowledge of J? ) computed by: GPOMDP [GPOMDP-J? ]; with baseline Ar = [J? ] [BL- [J? ]]; with optimum baseline [BL-A?r (y)]; with optimum constant baseline [BL-A?r ]; with pre-trained fixed value function [FVF-pretrain]. Online plots do a similar comparison of estimates computed by: GPOMDP [GPOMDP]; with online trained baseline [BL-online]; with online trained fixed value function [FVFonline]. The plots were computed over 500 runs (1000 for FVF-online), with ? = 0.95. ?? /T was set to 0.001 for FVF-pretrain, and 0.01 for FVF-online.   [7] P. W. Glynn. Likelihood ratio gradient estimation for stochastic systems. Communications of the ACM, 33:75?84, 1990. [8] E. Greensmith, P. L. Bartlett, and J. Baxter. Variance reduction techniques for gradient estimates in reinforcement learning. Technical report, ANU, 2002. [9] H. Kimura, K. Miyazaki, and S. Kobayashi. Reinforcement learning in POMDPs with function approximation. In D. H. Fisher, editor, Proceedings of the Fourteenth International Conference on Machine Learning (ICML?97), pages 152?160, 1997. [10] V. R. Konda and J. N. Tsitsiklis. Actor-Critic Algorithms. In Advances in Neural Information Processing Systems 12, pages 1008?1014. MIT Press, 2000. [11] P. Marbach and J. N. Tsitsiklis. Simulation-Based Optimization of Markov Reward Processes. Technical report, MIT, 1998. [12] R. Y. Rubinstein. How to optimize complex stochastic systems from a single sample path by the score function method. Ann. Oper. Res., 27:175?211, 1991. [13] R. S. Sutton and A. G. Barto. Reinforcement Learning: An Introduction. MIT Press, Cambridge MA, 1998. ISBN 0-262-19398-1. [14] R. S. Sutton, D. McAllester, S. Singh, and Y. Mansour. Policy Gradient Methods for Reinforcement Learning with Function Approximation. In Advances in Neural Information Processing Systems 12, pages 1057?1063. MIT Press, 2000. [15] R. J. Williams. Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning. Machine Learning, 8:229?256, 1992.
2052 |@word briefly:1 version:1 norm:6 simulation:1 covariance:1 reduction:3 initial:2 score:1 selecting:3 current:1 com:1 comparing:1 analysed:1 readily:1 evans:1 additive:2 plot:5 stationary:4 intelligence:2 xk:2 ith:1 short:1 consists:1 x0:7 indeed:2 expected:6 discounted:5 decreasing:1 csl:1 considering:1 becomes:1 notation:1 bounded:5 null:1 miyazaki:1 pto:1 minimizes:2 z:6 whilst:1 finding:1 kimura:1 temporal:1 control:6 normally:1 omit:2 appear:1 producing:1 greensmith:2 positive:1 t1:1 engineering:1 dropped:1 understood:1 before:1 limit:1 kobayashi:1 sutton:3 oxford:1 path:7 au:2 smc:1 unique:1 evan:2 whizbang:2 pre:2 wait:1 suggest:2 cal:1 py:2 optimize:1 deterministic:1 yt:16 williams:1 starting:1 pomdp:20 ergodic:3 simplicity:1 notion:1 variation:1 suppose:2 exact:6 us:3 element:1 satisfying:6 environment:2 reward:11 trained:8 depend:3 singh:1 upon:1 easily:1 joint:1 describe:1 monte:3 artificial:2 rubinstein:1 choosing:1 widely:2 solve:1 ability:1 cov:4 analyse:1 final:1 online:16 obviously:1 eigenvalue:1 isbn:1 mixing:5 optimum:7 r1:1 produce:4 generating:1 derive:1 illustrate:1 school:1 implies:2 australian:2 uu:2 concentrate:1 stochastic:2 mcallester:1 behaviour:1 preliminary:3 exp:2 mapping:1 a2:1 estimation:4 largest:1 tool:1 weighted:3 mit:5 always:2 aim:2 barto:2 improvement:7 likelihood:1 pretrain:4 baseline:48 dependent:2 a0:4 equal:2 construct:1 icml:1 nearly:1 pxt:1 report:2 connectionist:1 national:2 replaced:5 neuronlike:1 investigate:1 chain:7 integral:5 euclidean:1 re:1 column:1 ar:25 deviation:3 international:1 ie:2 together:1 squared:7 settled:1 choose:2 derivative:2 oper:1 alteration:2 b2:1 baird:1 depends:1 performed:1 lab:1 start:1 actorcritic:1 minimize:2 variance:37 produced:1 iid:4 comparably:1 carlo:3 pomdps:1 cybernetics:1 glynn:1 obvious:1 proof:4 associated:1 popular:1 knowledge:1 ut:14 probp:1 attained:1 dt:3 rand:1 though:1 strongly:1 anderson:1 furthermore:3 until:2 correlation:1 hand:2 d:3 replacing:1 reversible:1 defines:1 perhaps:1 mdp:1 effect:1 true:10 unbiased:5 hence:1 moore:1 rl:4 exponentially:1 extend:1 slight:1 cambridge:1 marbach:1 access:1 actor:3 pu:1 certain:1 ofr:1 seen:3 minimum:1 additional:1 greater:1 converge:2 maximize:2 swartz:1 ii:1 relates:1 full:1 reduces:1 d0:2 technical:2 visit:1 y:16 controlled:13 involving:1 variant:2 controller:1 expectation:3 limt:2 background:1 rest:1 ascent:2 tend:1 gpomdp:16 call:1 easy:2 baxter:5 variate:2 suboptimal:3 reduce:1 idea:1 bartlett:6 suffer:1 peter:2 action:1 generally:3 clear:1 extensively:1 reduced:1 exist:1 estimated:2 write:1 sum:2 run:1 fourteenth:1 parameterized:7 throughout:2 decision:2 def:11 bound:12 quadratic:1 min:3 performing:1 according:1 describes:1 modification:1 ln:1 needed:1 know:1 end:2 generic:1 denotes:6 konda:1 restrictive:1 approximating:1 bl:12 added:1 quantity:4 p2t:1 dependence:1 biowulf:1 rt:3 gradient:31 distance:6 trivial:1 length:2 ratio:1 minimizing:3 difficult:1 relate:1 policy:6 upper:1 observation:7 markov:11 finite:4 descent:2 extended:1 communication:1 mansour:1 arbitrary:2 pair:1 usually:1 max:1 including:1 suitable:1 event:1 weaver:1 technology:1 auto:1 relative:5 var:14 pij:3 editor:1 uncorrelated:1 critic:3 pi:2 row:1 surprisingly:2 transpose:1 truncation:1 tsitsiklis:2 bias:2 side:2 distributed:1 transition:2 valid:1 author:1 reinforcement:10 adaptive:1 transaction:1 approximate:2 observable:2 assumed:1 search:2 learn:2 investigated:1 complex:1 main:1 arise:1 wish:1 rk:8 down:1 theorem:16 z0:4 xt:19 jt:5 showing:1 x:2 exists:3 adding:1 magnitude:1 anu:3 horizon:2 expressed:2 partially:2 satisfies:2 acm:1 ma:1 ann:1 man:1 fisher:1 infinite:4 reducing:1 uniformly:3 lemma:5 total:1 experimental:1 est:1 east:1 select:3 arises:1 jonathan:1 reactive:2 correlated:1
1,155
2,053
Rates of Convergence of Performance Gradient Estimates Using Function Approximation and Bias in Reinforcement Learning Gregory Z. Grudic University of Colorado, Boulder grudic@cs.colorado.edu Lyle H. Ungar University of Pennsylvania ungar@cis.upenn.edu Abstract We address two open theoretical questions in Policy Gradient Reinforcement Learning. The first concerns the efficacy of using function approximation to represent the state action value function, . Theory is presented showing that linear function approximation representations of can degrade the rate of convergence of performance gradient estimates by a factor of relative to when no function approximation of is used, where is the number of possible actions and is the number of basis functions in the function approximation representation. The second concerns the use of a bias term in estimating the state action value function. Theory is presented showing that a non-zero bias term can improve the rate of convergence of performance gradient estimates by , where is the number of possible actions. Experimental evidence is presented showing that these theoretical results lead to significant improvement in the convergence properties of Policy Gradient Reinforcement Learning algorithms.         1 Introduction Policy Gradient Reinforcement Learning (PGRL) algorithms have recently received attention because of their potential usefulness in addressing large continuous reinforcement Learning (RL) problems. However, there is still no widespread agreement on how PGRL algorithms should be implemented. In PGRL, the agent?s policy is characterized by a set of parameters which in turn implies a parameterization of the agent?s performance metric. Thus if represents a dimensional parameterization of the agent?s policy and is a performance metric the agent is meant to maximize, then the performance metric must have the form [6]. PGRL algorithms work by first estimating the performance gradient (PG) and then using this gradient to update the agent?s policy using:    !"#!   /  ! (1) $&%('*)+$-,. !0 where . is a small positive step size. If the estimate of !"#! is accurate, then the agent can climb the performance gradient in the  parameter space, toward locally optimal policies. In practice, !12! is estimated using samples of the state action value function . The PGRL formulation is attractive because 1) the parameterization  of the policy can directly imply  a generalization over the agent?s state space (e.g., can represent the adjustable weights in a neural network approximation), which suggests that PGRL algorithms can work well is on very high dimensional problems [3]; 2) the computational cost of estimating linear in the number of parameters , which contrasts with the computational cost for most RL algorithms which grows exponentially with the dimension of the state space; and 3) PG [6, 5, 4, 2, 1]. algorithms exist which are guaranteed to give unbiased estimates of !0"#!  !0"#! This paper addresses two open theoretical questions in PGRL formulations. In PGRL formulations performance gradient estimates typically have the following form: / 0!  )  #' 1'    #'          (2) ! where   !" #! is the estimate of the value of executing action $! in state ! (i.e. the state action value function), # %!  the bias subtracted from   !& & #! in state ! , ' is the number of steps the agent takes before estimating !12!0 , and the form of the function    depends on the PGRL algorithm being used (see Section 2, equation (3) for the form being considered here). The effectiveness of PGRL algorithms strongly depends on how   !" & #!  is obtained and the form of # %!  . The aim of this work is to address these questions. The first open theoretical question addressed here is concerned with the use of function approximation (FA) to represent the state action value function , which is in turn used to estimate the performance gradient. The original formulation of PGRL [6], the REINFORCE algorithm, has been largely ignored because of the slow rate of convergence of the PG estimate. The use of FA techniques to represent based on its observations has been suggested as a way of improving convergence properties. It has been proven that specific linear FA formulations can be incorporated into PGRL algorithms, while still guaranteeing convergence to locally optimal solutions [5, 4]. However, whether linear FA representations actually improves the convergence properties of PGRL is an open question. We present theory showing that using linear basis function representations of , rather than direct observations of it, can slow the rate of convergence of PG estimates by a factor of (see Theorem 1 in Section 3.1). This result suggests that PGRL formulations should avoid the use of linear FA techniques to represent . In Section 4, experimental evidence is presented supporting this conjecture.   #  The second open theoretical question addressed here is can a non-zero bias term in (2) improve the convergence properties of PG estimates? There has been speculation that an appropriate choice of can improve convergence properties [6, 5], but theoretical support has been lacking. This paper presents theory showing that if , where is the number actions, then the rate of convergence of the PG estimate is improved by (see Theorem 2 in Section 3.2). This suggests that the convergence properties of PGRL algorithms can be improved by using a bias term that is the average of values in each state. Section 4 gives experimental evidence supporting this conjecture.  #)(*  + & 1  #     # #  ) 2 The RL Formulation and Assumptions , . - / 0$ 12 :9 + & $ )F G2  $ 43 365 87 $ <; ) ' 1 #= * @A> $ )B ?C >+-  $&%(' )D A&E  $ ? " @ H @* )JIK-> $&%(' E  $ )L+ & $ )MG 2 NO+  A 3P & ;     Q  + & R   )S?C>+-% $ )D E  $ )S+ RT2 NO U3P & U; The RL problem is modeled as a Markov Decision Process (MDP). The agent?s state at time is given by , . At each time step the agent chooses from a finite set of actions and receives a reward . The dynamics of the environment are characterized by transition probabilities and expected rewards , . The policy followed by the agent is characterized by a parameter vector , and is defined by the probability distribution , . We ) assume that Q  + & R  is differentiable with respect to  .  Q    + & 1     Q  )I $(  ' $ > $  / &Q   +  )I  ( ('  ' > $&%   $ ) + & $ )  "Q exact expression for the performance gradient is: where  the  . !Then  )      = !GQ + & ! R     + & !      (3) !  ! @  !  ( '  E where     ) ( $ $ -  $ )  "Q2 and #     .   , under This policy gradient formulation requires that the state-action value function,  the! current policy be estimated.! This estimate,  , is derived using the observed value @  + & !  . We assume that ! @  + & !  has the following form: @  / #!  )  + & #!  ,#"  + & #!   where  "  + # !  has zero mean and finite variance $&@!% ' *)( . Therefore, if   + !  is an es !  timate of  / #!  obtained by averaging * observations of @  + & #!  , then the mean and variance are given by: (4) IB    / #! ! )   + & #! O ,+B    + & #!  ).-013/5 2 4 ( @  / !  are independently distributed. This is consistent In addition, we assume that We use the Policy Gradient Theorem of Sutton et al. [5] and limit our analysis to the start and state action value state discount reward formulation. Here the reward function function are defined as: with the MDP assumption. 3 Rate of Convergence Results $679% 8;: ) !@ @AB' ! <>@C =0' ? 'EDEDEDE' =GF $6@!% ' * ( H$679% I J ) @N@A' ! @<GC KM' L 'EDEDEDE' =>F $6@!% ' * ( where $ @!% ' * ( is defined in (4) and O 79I J )QP (    % ( = SR BT @!' W*)(VU W!X  %ZY $ 79% I J @ !M(' O 798;: )QP (    % ( =  R BT @!R ' W* ( U W!X  %ZY $679% 8;: @ !M(' R Before stating the convergence theorems, we define the following: (5) (6) 3.1 Rate of Convergence of PIFA Algorithms  ]`_ ]    / #!  ) *) (   )\] [ *)( ' *)( '   (7)  '&^ _ * (3' ] ]* (3' where ]  are weights and    are basis functions defined in  + 7 . If the weights _ *)( ' ] ^ ^ * (a' are chosen based using the observed ! @  + & !  , and the basis functions,    , satisfy the conditions defined in [5,= 4], then the performance gradient is given by: ! )     !Q + & ! R   *  (  (8) !cb ! @  ! ( ' The following theorem establishes bounds on the rate of convergence for this representation Consider the PIFA algorithm [5] which uses a basis function representation for estimated state action value function, , of the following form: of the performance gradient. RR  W b Theorem 1: Let be an estimate of (8) obtained using the PIFA algorithm and the basis function representation (7). Then, given the assumptions defined in Section 2 and equations (5) and (6), the rate of convergence of a PIFA  algorithm is bounded below and above by: O 97 I J  + ! O 798 :    *  !  b  *   where is the number of basis functions, is the number of possible actions, and the number of independent estimates of the performance gradient. * (9) is Proof: See Appendix. 3.2 Rate of Convergence of Direct Sampling Algorithms ! @  / #!   + & !  In the previous section, the observed are used to build a linear basis function representation of the state action value function, , which is in turn used to estimate the performance gradient. In this section we establish rate of convergence bounds for performance gradient estimates that directly use the observed without the intermediate step of building the FA representation. These bounds are established for the conditions and in (3). #  #  )  )  # / RR W ( *  / 1  @  + & #!  #   )   Theorem 2: Let be a estimate of (3), be obtained using direct samples of . Then, , and given the assumptions defined in Section 2 and equations (5) and (6), the if rate of convergence of * / RR W O 97 I J +  !0/  O 97 8;: *  !   * is bounded by: (10) where is the number of independent estimates of the performance gradient. If is defined as:   )    =   + &   ' then the rate of convergence of the performance gradient  where  # )   (11) RR  W is bounded by: O 97 I J  + ! O 97 8 :  *   0!    *  (12) is the number of possible actions. Proof: See Appendix. Thus comparing (12) and (10) to (9) one can see that policy gradient algorithms such as PIFA which build FA representations of converge by a factor of slower than algorithms which directly sample . Furthermore, if the bias term is as defined in (11), the bounds on the variance are further reduced by . In the next section experimental evidence is given showing that these theoretical consideration can be used to improve the convergence properties of PGRL algorithms.        4 Experiments The Simulated Environment: The experiments simulate an agent episodically interacting in a continuous two dimensional environment. The agent starts each episode in the same state  , and executes a finite number of steps following a policy to a fixed goal state  . The stochastic policy is defined by a finite set of Gaussians, each associated with a specific ! 0.6 4 2 10 b V[? ? / ? ?] / V[? ? / ? ? ] No Bias 0.4 3 10 0.3 Linear FA Q 0.2 2 10 0.1 ?0.1 0 1 10 0 20 40 60 80 100 0 10 Number of Policy Updates a) Convergence of Algorithms 0 1 10 F ?(?) 10 Biased Q V[? ? / ? ? ] / V[? ? / ? ?] 0.5 / / 0 2 4 6 8 10 12 Number of Possible Actions (M) b)   14 + RR W b  + RR W  10 0 / / 2 4 6 8 10 12 Number of Possible Actions (M) c) Figure 1: Simulation Results   + RR W   + RR W  14  is defined as:    %     ) Z?  7         ( ' where  )  #' 1    87 , is the agents state, *'%   7 7 is the Gaussian center, and  *' 1   is the variance along each state space dimension. The probability of 7  in state  is executing action   Q  + &  R   ) =        ((  ' where  )  '' 1 '  ' ' 1  ' 1 1= ' 1 1=   7  = '   = 7  defines the policy parameters that dictate the agent?s actions. Action ' directs the agent toward the goals state   , while the remaining actions   (for  ) 0T   ) direct the agent towards the corresponding Gaussian center *'%   . 7 Noise is modeled using a uniform random distribution between  $  denoted by  $  , such that the noise in dimension!  @ is given by:    )   ,   @ `)  #$ 2 where  9  is the magnitude of the noise,  is the state the agent observes and uses to  agent. choose actions, and  is the actual state of the  action. The Gaussian associated with action  The agent receives a reward of +1 when it reaches the goal state, otherwise it receives a reward of:  >  ) )  Z?  7  %   ( ' Thus the agent gets negative rewards the closer it gets to the origin of the state space, and a positive reward whenever it reaches the goal state.  @  + & !  Implementation of the PGRL algorithms: All the PGRL formulations studied here require observations (i.e. samples) of the state action value function. is sampled by executing action in state and thereafter following the policy. In the episodic formulation, where the agent executes a maximum of steps during each episode, at the end of each episode, for step can be evaluated as follows: T!   @   $ & $   @   $ & $   @   ' & '  1  @   Q &  ' ,    )   ' > $&%  E  $  ) + $ )  &Q ( '   ' 1'     #( Thus, given that the agent executes a complete episode following the policy , at the completion of the episode we can calculate . This gives samples of state action value pairs. Equation (3) tells us that we require a total of state action value function observations to estimate a performance gradient (assuming the agent can execute actions). Therefore, we can obtain the remaining observations of by sending the agent out on    + '  ' '  @   ' $ U)  ! @   $ &   Q  '   ' & '  1      ' epsisodes, each time allowing it to follow the policy for all steps, with the exception that action is executed when is being observed. This sampling procedure requires a total of episodes and gives a complete set of state action pairs for any path . For the direct sampling algorithms in Section 3.2, these observations are directly used to estimate the performance gradient. For the linear basis function based PGRL algorithm in Section 3.1, these observations are as defined in [5, 4], and then the performance gradient is first used to calculate the calculated using (8). ^ * ( '] /  @ / + !"#! b   + !12!0  Experimental Results: Figure 1b shows a plot of average values over 10,000 estimates of the performance gradient. For each estimate, the goal state, start ; state, and Gaussian centers are all chosen using a uniform random distribution the Gaussian variances are sampled from a uniform distribution . As predicted by Theorem 1 in Section 3.1 and Theorem 2 in Section 3.2, as the number of actions increases, this ratio also increases. Note that Figure 1b plots average variance ratios, not the bounds in variance given in Theorem 1 and Theorem 2 (which have not been experimentally sampled), so the ratio predicted by the theorems is supported by the increase in increases. Figure 1c shows a plot of average values the ratio as over 10,000 estimates of the performance gradient. As above, for each estimate, the goal state, start state, and Gaussian centers are all chosen using a uniform random distribution ; the Gaussian variances are sampled from a uniform distribution . This also follows the predicted trends of Theorem 1 and Theorem 2. Finally, Figure 1a shows the average reward over 100 runs as the three algorithms converge on a two action problem. Each algorithm is given the same number of samples to estimate the gradient before each update. Because has the least variance, it allows the policy to converge . Similarly, because has the highest variance, its to the highest reward value policy updates converge to the worst . Note that because all three algorithms will converge to the same locally optimal policy given enough samples of , Figure 1a simply requires more samples than , which in turn requires more demonstrates that samples than .  +    $ /     +  / 0! "#!  Q  /  ! / 12! !"#! b ! @ Q  + !/ 12!0   + !/ "#!  `) +  / ! 12!0 b / ! 1#! Q ! @ 5 Conclusion The theoretical and experimental results presented here indicate that how PGRL algorithms are implemented can substantially affect the number of observations of the state action value function ( ) needed to obtain good estimates of the performance gradient. Furthermore, they suggest that an appropriately chosen bias term, specifically the average value of over all actions, and the direct use of observed values can improve the convergence of PGRL algorithms. In practice linear basis function representations of can significantly degrade the convergence properties of policy gradient algorithms. This leaves open the question of whether any (i.e. nonlinear) function approximation representation of value functions can be used to improve convergence of such algorithms. References [1] Jonathan Baxter and Peter L. Bartlett, Reinforcement learning in pomdp?s via direct gradient ascent, Proceedings of the Seventeenth International Conference on Machine Learning (ICML?2000) (Stanford University, CA), June 2000, pp. 41?48. [2] G. Z. Grudic and L. H. Ungar, Localizing policy gradient estimates to action transitions, Proceedings of the Seventeenth International Conference on Machine Learning, vol. 17, Morgan Kaufmann, June 29 - July 2 2000, pp. 343?350. [3] , Localizing search in reinforcement learning, Proceedings of the Seventeenth National Conference on Artificial Intelligence, vol. 17, Menlo Park, CA: AAAI Press / Cambridge, MA: MIT Press, July 30 - August 3 2000, pp. 590?595. [4] V. R. Konda and J. N. Tsitsiklis, Actor-critic algorithms, Advances in Neural Information Processing Systems (Cambridge, MA) (S. A. Solla, T. K. Leen, and K.-R. Mller, eds.), vol. 12, MIT Press, 2000. [5] R. S. Sutton, D. McAllester, S. Singh, and Y. Mansour, Policy gradient methods for reinforcement learning with function approximation, Advances in Neural Information Processing Systems (Cambridge, MA) (S. A. Solla, T. K. Leen, and K.-R. Mller, eds.), vol. 12, MIT Press, 2000. [6] R. J. Williams, Simple statistical gradient-following algorithms for connectionist reinforcement learning, Machine Learning 8 (1992), no. 3, 229?256. *  ( given in (7). In [5] it is shown that _ *)( ' ]   = !Q  + & ! R   ^  = !Q  + & ! R    I  @    !( ' ! * (    @      + & !   )  !M(' ! (13)  !  Let R W @ be the observation of R W (3) after a single episode. Using (13), we get the R R following: = N @ ' ( B  T *  ! RR W @ ) ( @     !( ( ' R R W U W!X ! @ / #!  RR W , " ) P (     ( = R BT @!' *Z( U W;X   + & !  Y , " @ !M= ' BT @!R ' W * ( U W;X ) P ( @     !M(  ' R R W *) (    Y ,#" = ]`_ ] ) P ( @    !M(  ' R BT @!R ' W * ( U W;X ] (  [ ' *)( ' *)( '   Y ,#" ^ _ ] = = ]  ] ] ) P !M( (' ] ( ([ ' *Z(  ' ] ( @ R BT @!R ' W * ( U W X      *Z( '    Y , " !M(  ' ] ( ([ ' *)( ' ! ,#" ^ ^ where the basis functions ! have the form   ] _ ] ! )  @ !GQ  /!  ! R      * ( '  I "  )  , with variance and  = !Q  + & ! R   %   !  Y %   + "  )+ P !  @ ) @      !(' ! $ @!% ' * (  Denoting R W b as the least squares (LS) estimate of (3), its form is given by: R ! ) = [   (14) !0 b  ( '  ] where   are] LS estimates of the  weights *)( ' and  correspond to the  basis  ! . Then, it can be shown that any linear ^ system of the type given in (14) has a functions rate of convergence given by:  = !GQ + & #!R   %   !        + ! b  ) * + "  ) *  @     % !( ' ! $ @!% ' * ( Substituting (5) and (6) into the above equation completes the proof. Appendix: Proofs of Theorems 1 and 2 *Z( ' ] Proof of Theorem 1: Consider the definition of there exist and such that:  of the performance ! @  / #* ! . estimates These examples are averaged / =  ! I !  )  @     !( ' !GQ  /!  ! R    / !  ! @  + & !  is independently distributed, the variance of the estimate is given Because each by / = !Q + & #!R  %   !  % + !  ) *  @     ! $ @N% ' *)( (15) !( ' Given (5) the worst / rate of convergence= is bounded by:  ! + !    @     % !( ' !Q  +! & ! R   %  $ 79% 8 :  * ) O 798 : * A similarly argument applies to the lower bound on convergence completing the proof for (10). Following the same argument for (12), we have  = !Q + & ! R  %  =    !  + !  ) *  @     % !( ' !  +   + !    ( '  +   Where  = = +   + & !  = '  ( ( '   + &    )+  = =  '  + & !  = '  ( ( '   /       ! =  ) =  = ' % $6@!% ' *)( , ( (  ' = ' % $6@N% ' * ! (16) +  on the far left of (16) is bounded by: Given (5) the variance  = =    =  = ' % $6@!% ' *)( , ( ( ' = ' % $6@!% ' *    ) =  = ' % $679% 8;: , ( ( ' = ' % $679% 8 : ! ! 798;:  = ' % )   = ' ,    = ' %  $679% 8;: O )  = $679% 8 : Plugging the above into (16) and inserting  *  from (6) completes the proof for the upper * Proof of Theorem 2: We prove equation (10) first. For gradient, we get independent samples of each and therefore:  bound. The proof for the lower bound in the variance follows similar reasoning.
2053 |@word build:2 implemented:2 c:1 predicted:3 implies:1 indicate:1 unbiased:1 establish:1 question:7 open:6 mdp:2 km:1 fa:8 simulation:1 stochastic:1 attractive:1 pg:6 mcallester:1 during:1 gradient:35 require:2 reinforce:1 simulated:1 ungar:3 generalization:1 efficacy:1 degrade:2 denoting:1 complete:2 toward:2 assuming:1 current:1 comparing:1 reasoning:1 considered:1 modeled:2 consideration:1 ratio:4 recently:1 must:1 cb:1 mller:2 executed:1 substituting:1 rl:4 qp:2 negative:1 plot:3 exponentially:1 update:4 implementation:1 intelligence:1 leaf:1 policy:26 adjustable:1 parameterization:3 significant:1 allowing:1 cambridge:3 upper:1 observation:10 markov:1 finite:4 establishes:1 similarly:2 supporting:2 mit:3 incorporated:1 gaussian:7 aim:1 gc:1 rather:1 interacting:1 avoid:1 actor:1 mansour:1 along:1 august:1 direct:7 prove:1 derived:1 june:2 pair:2 improvement:1 directs:1 speculation:1 contrast:1 upenn:1 expected:1 established:1 address:3 suggested:1 morgan:1 below:1 typically:1 bt:6 actual:1 converge:5 maximize:1 july:2 estimating:4 bounded:5 denoted:1 substantially:1 characterized:3 jik:1 improve:6 sampling:3 plugging:1 imply:1 represents:1 park:1 icml:1 metric:3 demonstrates:1 t2:1 connectionist:1 represent:5 gf:1 addition:1 positive:2 before:3 national:1 addressed:2 relative:1 completes:2 limit:1 lacking:1 sutton:2 appropriately:1 biased:1 proven:1 ab:1 sr:1 path:1 ascent:1 agent:25 studied:1 climb:1 suggests:3 effectiveness:1 consistent:1 critic:1 intermediate:1 enough:1 seventeenth:3 averaged:1 concerned:1 baxter:1 accurate:1 affect:1 supported:1 lyle:1 practice:2 vu:1 closer:1 pennsylvania:1 tsitsiklis:1 bias:9 procedure:1 episodic:1 whether:2 expression:1 distributed:2 significantly:1 dictate:1 theoretical:8 bartlett:1 dimension:3 calculated:1 transition:2 suggest:1 peter:1 get:4 reinforcement:9 localizing:2 far:1 action:37 cost:2 ignored:1 addressing:1 pgrl:21 uniform:5 usefulness:1 center:4 discount:1 williams:1 attention:1 locally:3 independently:2 l:2 pomdp:1 reduced:1 gregory:1 exist:2 chooses:1 continuous:2 search:1 international:2 estimated:3 ca:2 menlo:1 vol:4 improving:1 thereafter:1 colorado:2 exact:1 aaai:1 us:2 choose:1 origin:1 agreement:1 grows:1 trend:1 noise:3 potential:1 observed:6 run:1 worst:2 calculate:2 satisfy:1 slow:2 episode:7 depends:2 solla:2 highest:2 decision:1 observes:1 appendix:3 environment:3 start:4 bound:8 reward:10 grudic:3 guaranteed:1 followed:1 completing:1 dynamic:1 ib:1 theorem:17 singh:1 specific:2 square:1 showing:6 variance:14 largely:1 kaufmann:1 correspond:1 basis:12 concern:2 evidence:4 simulate:1 argument:2 zy:2 ci:1 magnitude:1 conjecture:2 executes:3 artificial:1 tell:1 reach:2 whenever:1 ed:2 simply:1 definition:1 stanford:1 pp:3 otherwise:1 g2:1 proof:9 associated:2 applies:1 timate:1 sampled:4 boulder:1 equation:6 differentiable:1 rr:10 mg:1 improves:1 turn:4 ma:3 goal:6 needed:1 gq:4 towards:1 actually:1 inserting:1 end:1 sending:1 experimentally:1 gaussians:1 follow:1 specifically:1 improved:2 averaging:1 formulation:11 evaluated:1 execute:1 strongly:1 leen:2 furthermore:2 appropriate:1 total:2 subtracted:1 experimental:6 e:1 convergence:30 slower:1 receives:3 exception:1 original:1 guaranteeing:1 executing:3 nonlinear:1 remaining:2 widespread:1 support:1 defines:1 stating:1 completion:1 meant:1 jonathan:1 konda:1 received:1 building:1
1,156
2,054
Adaptive N earest Neighbor Classification using Support Vector Machines Carlotta Domeniconi, Dimitrios Gunopulos Dept. of Computer Science, University of California, Riverside, CA 92521 { carlotta, dg} @cs.ucr.edu Abstract The nearest neighbor technique is a simple and appealing method to address classification problems. It relies on t he assumption of locally constant class conditional probabilities. This assumption becomes invalid in high dimensions with a finite number of examples due to the curse of dimensionality. We propose a technique that computes a locally flexible metric by means of Support Vector Machines (SVMs). The maximum margin boundary found by the SVM is used to determine the most discriminant direction over the query's neighborhood. Such direction provides a local weighting scheme for input features. We present experimental evidence of classification performance improvement over the SVM algorithm alone and over a variety of adaptive learning schemes, by using both simulated and real data sets. 1 Introduction In a classification problem, we are given J classes and l training observations. The training observations consist of n feature measurements x = (Xl,'" ,Xn)T E ~n and the known class labels j = 1, ... , J. The goal is to predict the class label of a given query q. The K nearest neighbor classification method [4, 13, 16] is a simple and appealing approach to this problem: it finds the K nearest neighbors of q in the training set, and then predicts the class label of q as the most frequent one occurring in the K neighbors. It has been shown [5, 8] that the one nearest neighbor rule has asymptotic error rate that is at most twice t he Bayes error rate, independent of the distance metric used. The nearest neighbor rule becomes less appealing with finite training samples, however. This is due to the curse of dimensionality [2]. Severe bias can be introduced in t he nearest neighbor rule in a high dimensional input feature space with finite samples. As such, the choice of a distance measure becomes crucial in determining t he outcome of nearest neighbor classification. The commonly used Euclidean distance implies that the input space is isotropic, which is often invalid and generally undesirable in many practical applications. Several techniques [9, 10, 7] have been proposed to try to minimize bias in high dimensions by using locally adaptive mechanisms. The "lazy learning" approach used by these methods, while appealing in many ways, requires a considerable amount of on-line computation, which makes it difficult for such techniques to scale up to large data sets. The feature weighting scheme they introduce, in fact , is query based and is applied on-line when the test point is presented to the "lazy learner" . In this paper we propose a locally adaptive metric classification method which, although still founded on a query based weighting mechanism, computes off-line the information relevant to define local weights. Our technique uses support vector machines (SVMs) as a guidance for the process of defining a local flexible metric. SVMs have been successfully used as a classification tool in a variety of areas [11, 3, 14], and the maximum margin boundary they provide has been proved to be optimal in a structural risk minimization sense. The solid theoretical foundations that have inspired SVMs convey desirable computational and learning theoretic properties to the SVM's learning algorithm, and therefore SVMs are a natural choice for seeking local discriminant directions between classes. The solution provided by SVMs allows to determine locations in input space where class conditional probabilities are likely to be not constant, and guides the extraction of local information in such areas. This process produces highly stretched neighborhoods along boundary directions when the query is close to the boundary. As a result, the class conditional probabilities tend to be constant in the modified neighborhoods, whereby better classification performance can be achieved. The amount of elongation-constriction decays as the query moves further from the boundary vicinity. 2 Feature Weighting SVMs classify patterns according to the sign(f(x)), where f(x) L:~=l (XiYiK(Xi, x) - b, K(x , y) = cpT(x). cp(y) (kernel junction), and cp: 3(n -+ 3(N is a mapping of the input vectors into a higher dimensional feature space. Here we assume Xi E 3(n, i = I, . . . ,l, and Yi E {-I,I}. Clearly, in the general case of a non-linear feature mapping cp, the SVM classifier gives a non-linear boundary f(x) = 0 in input space. The gradient vector lld = "Vdj, computed at any point d of the level curve f(x) = 0, gives the perpendicular direction to the decision boundary in input space at d. As such, the vector lld identifies the orientation in input space on which the projected training data are well separated, locally over d's neighborhood. Therefore, the orientation given by lld, and any orientation close to it, is highly informative for the classification task at hand , and we can use such information to define a local measure of feature relevance. Let q be a query point whose class label we want to predict. Suppose q is close to the boundary, which is where class conditional probabilities become locally non uniform, and therefore estimation of local feature relevance becomes crucial. Let d be the closest point to q on the boundary f(x) = 0: d = argminp Ilq - pll, subject to the constraint f(p) = O. Then we know that the gradient lld identifies a direction along which data points between classes are well separated. As a consequence, the subspace spanned by the orientation lld, locally at q, is likely to contain points having the same class label as q . Therefore, when applying a nearest neighbor rule at q, we desire to stay close to q along the lld direction, because that is where it is likely to find points similar to q in terms of class posterior probabilities. Distances should be constricted (large weight) along lld and along directions close to it. The farther we move from the lld direction, the less discriminant the correspondent orientation becomes. This means that class labels are likely not to change along those orientations, and distances should be elongated (small weight) , thus including in q's neighborhood points which are likely to be similar to q in terms of the class conditional probabilities. Formally, we can measure how close a direction t is to lld by considering the dot product lla ?t. In particular, by denoting with Uj the unit vector along input feature j, for j = 1, . .. , n, we can define a measure of relevance for feature j, locally at q (and therefore at d), as Rj(q) == Iu] . lldl = Ind,j l, where lld = (nd,l,'" ,nd,n)T. The measure of feature relevance, as a weighting scheme, can then be given by the following exponential weighting scheme: Wj(q) = exp(ARj(q))1 2::7=1 exp(ARi(q)), where A is a parameter that can be chosen to maximize (minimize) the influence of R j on Wj' When A = 0 we have Wj = lin, thereby ignoring any difference between the Rj's. On the other hand, when A is large a change in R j will be exponentially reflected in Wj' The exponential weighting scheme conveys stability to the method by preventing neighborhoods to extend infinitely in any direction. This is achieved by avoiding zero weights, which would instead be allowed by linear or quadratic weightings. Thus, the exponential weighting scheme can be used as weights associated with features for weighted distance computation D(x, y) = )2::7=1 Wi(Xi - Yi)2. These weights enable the neighborhood to elongate less important feature dimensions, and, at the same time, to constrict the most influential ones. Note that the technique is query-based because weightings depend on the query. 3 Local Flexible Metric Classification based on SVMs To estimate the orientation of local boundaries, we move from the query point along the input axes at distances proportional to a given small step (whose initial value can be arbitrarily small, and doubled at each iteration till the boundary is crossed). We stop as soon as the boundary is crossed along an input axis i, i.e. when a point P i is reached that satisfies the condition sign(f(q)) x sign(f(pi)) = -1. Given Pi, we can get arbitrarily close to the boundary by moving at (arbitrarily) small steps along the segment that joins Pi to q. Let us denote with d i the intercepted point on the boundary along direction i. We then approximate lld with the gradient vector lld i = \7 d i f, computed at d i . We desire that the parameter A in the exponential weighting scheme increases as the distance of q from the boundary decreases. By using the knowledge that support vectors are mostly located around the boundary surface, we can estimate how close a query point q is to the boundary by computing its distance from the closest non bounded support vector: Bq = minsi Ilq - si ll, where the minimum is taken over the non bounded (0 < D:i < C) support vectors Si. Following the same principle, in [1] the spatial resolution around the boundary is increased by enlarging volume elements locally in neighborhoods of support vectors. Then, we can achieve our goal by setting A = D - B q , where D is a constant input parameter of the algorithm. In our experiments we set D equal to the approximated average distance between the training points Xk and the boundary: D = 2::xk {minsi Ilxk - sill}. If A becomes negative it is set to zero. t By doing so the value of A nicely adapts to each query point according to its location with respect to the boundary. The closer q is to the decision boundary, the higher the effect of the Rj's values will be on distances computation. We observe that this principled guideline for setting the parameters of our technique takes advantage of the sparseness representation of the solution provided by the SVM. In fact, for each query point q, in order to compute Bq we only need to consider the support vectors, whose number is typically small compared to the Input: Decision boundary f(x) point q and parameter K. = a produced by a SVM; query 1. Compute the approximated closest point d i to q on the bound- ary; 2. Compute the gradient vector ndi = \l dJ; 3. Set feature relevance values Rj(q) = Indi,jl for j = 1, . . . ,n; 4. Estimate the distance of q from the boundary as: Bq = minsi Ilq - sill; 5. Set A = D - B q , where D = EXk {minsi Ilxk - sill}; t 6. Set Wj(q) = exp(ARj(q))/ E~=l exp(ARi(q)), for j 1, ... ,n; 7. Use the resulting w for K-nearest neighbor classification at the query point q. Figure 1: The LFM-SVM algorithm total number of training examples. Furthermore, the computation of D's value is carried out once and off-line. The resulting local flexible metric technique based on SVMs (LFM-SVM) is summarized in Figure 1. The algorithm has only one adjustable tuning parameter, namely the number K of neighbors in the final nearest neighbor rule. This parameter is common to all nearest neighbor classification techniques. 4 Experimental Results In the following we compare several classification methods using both simulated and real data. We compare the following classification approaches: (1) LFM-SVM algorithm described in Figure 1. SV Mlight [12] with radial basis kernels is used to build the SVM classifier; (2) RBF-SVM classifier with radial basis kernels. We used SV Mlight [12], and set the value of"( in K(Xi' x) = e-r llxi-xI12 equal to the optimal one determined via cross-validation. Also the value of C for the soft-margin classifier is optimized via cross-validation. The output of this classifier is the input of LFM-SVM; (3) ADAMENN-adaptive metric nearest neighbor technique [7]. It uses the Chi-squared distance in order to estimate to which extent each dimension can be relied on to predict class posterior probabilities; (4) Machete [9]. It is a recursive partitioning procedure, in which the input variable used for splitting at each step is the one that maximizes the estimated local relevance. Such relevance is measured in terms of the improvement in squared prediction error each feature is capable to provide; (5) Scythe [9]. It is a generalization of the machete algorithm, in which the input variables influence each split in proportion to their estimated local relevance; (6) DANN-discriminant adaptive nearest neighbor classification [10]. It is an adaptive nearest neighbor classification method based on linear discriminant analysis. It computes a distance metric as a product of properly weighted within and between sum of squares matrices; (7) Simple K-NN method using the Euclidean distance measure; (8) C4.5 decision tree method [15]. In all the experiments, the features are first normalized over the training data to have zero mean and unit variance, and the test data features are normalized using the corresponding training mean and variance. Procedural parameters (including K) for each method were determined empirically through cross-validation. 4.1 Experiments on Simulated Data For all simulated data, 10 independent training samples of size 200 were generated. For each of these, an additional independent test sample consisting of 200 observations was generated. These test data were classified by each competing method using the respective training data set. Error rates computed over all 2,000 such classifications are reported in Table 1. The Problems. (1) Multi-Gaussians. The data set consists of n = 2 input features, l = 200 training data, and J = 2 classes. Each class contains two spherical bivariate normal subclasses, having standard deviation 1. The mean vectors for one class are (-3/4, -3) and (3/4,3); whereas for the other class are (3, -3) and (-3,3). For each class, data are evenly drawn from each of the two normal subclasses. The first column of Table 1 shows the results for this problem. The standard deviations are: 0.17, 0.01, 0.01, 0.01, 0.01 0.01, 0.01 and 1.50, respectively. (2) Noisy-Gaussians. The data for this problem are generated as in the previous example, but augmented with four predictors having independent standard Gaussian distributions. They serve as noise. Results are shown in the second column of Table 1. The standard deviations are: 0.18, 0.01, 0.02, 0.01, 0.01, 0.01, 0.01 and 1.60, respectively. Results. Table 1 shows that all methods have similar performances for the MultiGaussians problem, with C4.5 being the worst performer. When the noisy predictors are added to the problem (NoisyGaussians), we observe different levels of deterioration in performance among the eight methods. LFM-SVM shows the most robust behavior in presence of noise. K-NN is instead the worst performer. In Figure 2 we plot the performances of LFM-SVM and RBF-SVM as a function of an increasing number of noisy features (for the same MultiGaussians problem). The standard deviations for RBF-SVM (in order of increasing number of noisy features) are: 0.01, 0.01 , 0.03, 0.03, 0.03 and 0.03. The standard deviations for LFM-SVM are: 0.17,0.18,0.2,0.3,0.3 and 0.3. The LFM-SVM technique shows a considerable improvement over RBF -SVM as the amount of noise increases. Table 1: Average classification error rates for simulated and real data. LFM-SVM RBF-SVM ADAMENN Machete Scythe DANN K-NN C4.5 4.2 MultiGauss NoisyGauss Iris Sonar Liver Vote Breast OQ Pima 3.3 3.4 4.0 11.0 28.1 2.6 3.0 3.5 19.3 3.3 4.1 4.0 12.0 26.1 3.0 3.1 3.4 21.3 3.4 4.1 3.0 9.1 30.7 3.0 3.2 3.1 20.4 3.4 4.3 5.0 21.2 27.5 3.4 3.5 7.4 20.4 4.0 16.3 27.5 3.4 4.8 2.7 3.4 5.0 20.0 2.2 4.0 22.2 4.7 3.7 6.0 1.1 30.1 3.0 3.3 7.0 6.0 12.5 32.5 7.8 2.7 5.4 24.2 5.0 5.1 8.0 23.1 38.3 3.4 4.1 9.2 23.8 Experiments on Real Data In our experiments we used seven different real data sets. They are all taken from DCI Machine Learning Repository at http://www.cs.uci.edu/,,,-,mlearn/ MLRepository.html. For a description of the data sets see [6]. For the Iris, Sonar, Liver and Vote data we perform leave-one-out cross-validation to measure performance, since the number of available data is limited for these data sets. For the 36'--'--'---r--'--~--'--'--~--.--'--~ LFM-SVM --+-RBF-SVM ---)(--- 34 32 30 28 26 24 22 20 18 16 14 12 10 8 6 ~ ~~=='P'O L-~--~--~~--~--~~--~--~~--~ o 10 12 14 16 18 20 22 Number of Noisy Variables Figure 2: Average Error Rates of LFM-SVM and RBF-SVM as a function of an increasing number of noisy predictors. I I J. T I - -? -? - 1- ""'!"" :E ~ ;l "" ..J :E > :z ""' ~ z zOJ :E "" "" Q i ~ " 1j ~ i z z "" Q z z '" 3 Figure 3: Performance distributions for real data. Breast, OQ-Ietter and Pima data we randomly generated five independent training sets of size 200. For each of these, an additional independent test sample consisting of 200 observations was generated. Table 1 (columns 3-9) shows the cross-validated error rates for the eight methods under consideration on the seven real data. The standard deviation values are as follows. Breast data: 0.2, 0.2, 0.2, 0.2, 0.2, 0.9, 0.9 and 0.9, respectively. OQ data: 0.2 , 0.2 , 0.2, 0.3, 0.2 , 1.1 , 1.5 and 2.1 , respectively. Pima data: 0.4, 0.4, 0.4, 0.4, 0.4, 2.4, 2.1 and 0.7, respectively. Results. Table 1 shows that LFM-SVM achieves the best performance in 2/7 of the real data sets; in one case it shows the second best performance, and in the remaining four its error rate is still quite close to t he best one. Following Friedman [9], we capture robustness by computing the ratio bm of the error rate em of method m and the smallest error rate over all methods being compared in a particular example: bm = emf minl~k~8 ek? Figure 3 plots the distribution of bm for each method over the seven real data sets. The dark area represents the lower and upper quartiles of the distribution that are separated by the median. The outer vertical lines show the entire range of values for the distribution. The spread of the error distribution for LFM-SVM is narrow and close to one. The results clearly demonstrate that LFM-SVM (and ADAMENN) obtained the most robust performance over the data sets. The poor performance of the machete and C4.5 methods might be due to the greedy strategy they employ. Such recursive peeling strategy removes at each step a subset of data points permanently from further consideration. As a result, changes in an early split, due to any variability in parameter estimates, can have a significant impact on later splits , thereby producing different terminal regions. This makes predictions highly sensitive to the sampling fluctuations associated with the random nature of the process that produces the traning data, thus leading to high variance predictions. The scythe algorithm, by relaxing the winner-take-all splitting strategy of the machete algorithm, mitigates the greedy nature of the approach, and thereby achieves better performance. In [10], the authors show that the metric employed by the DANN algorithm approximates the weighted Chi-squared distance, given that class densities are Gaussian and have the same covariance matrix. As a consequence, we may expect a degradation in performance when the data do not follow Gaussian distributions and are corrupted by noise , which is likely the case in real scenarios like the ones tested here. We observe that the sparse solution given by SVMs provides LFM-SVM with principled guidelines to efficiently set the input parameters. This is an important advantage over ADAMENN, which has six tunable input parameters. Furthermore, LFM-SVM speeds up the classification process since it applies the nearest neighbor rule only once, whereas ADAMENN applies it at each point within a region centered at the query. We also observe that the construction of the SVM for LFM-SVM is carried out off-line only once, and there exist algorithmic and computational results which make SVM training practical also for large-scale problems [12]. The LFM-SVM offers performance improvements over the RBF-SVM algorithm alone, for both the (noisy) simulated and real data sets. The reason for such performance gain may rely on the effect of our local weighting scheme on the separability of classes, and therefore on the margin, as shown in [6]. Assigning large weights to input features close to the gradient direction, locally in neighborhoods of support vectors, corresponds to increase the spatial resolution along those orientations, and therefore to improve the separability of classes. As a consequence, better classification results can be achieved as demonstrated in our experiments. 5 Related Work In [1], Amari and Wu improve support vector machine classifiers by modifying kernel functions. A primary kernel is first used to obtain support vectors. The kernel is then modified in a data dependent way by using the support vectors: the factor that drives the transformation has larger values at positions close to support vectors. The modified kernel enlarges the spatial resolution around the boundary so that the separability of classes is increased. The resulting transformation depends on the distance of data points from the support vectors , and it is therefore a local transformation, but is independent of the boundary's orientation in input space. Likewise, our transformation metric depends , through the factor A, on the distance of the query point from the support vectors. Moreover, since we weight features, our metric is directional, and depends on the orientation of local boundaries in input space. This dependence is driven by our measure of feature relevance, which has the effect of increasing the spatial resolution along discriminant directions around the boundary. 6 Conclusions We have described a locally adaptive metric classification method and demonstrated its efficacy through experimental results. The proposed technique offers performance improvements over the SVM alone, and has the potential of scaling up to large data sets. It speeds up, in fact, the classification process by computing offline the information relevant to define local weights, and by applying the nearest neighbor rule only once. Acknowledgments This research has been supported by the National Science Foundation under grants NSF CAREER Award 9984729 and NSF IIS-9907477, by the US Department of Defense, and a research award from AT&T. References [1] S. Amari and S. Wu, "Improving support vector machine classifiers by modifying kernel functions", Neural Networks, 12, pp. 783-789, 1999. [2] R.E. Bellman, Adaptive Control Processes. Princeton Univ. Press, 1961. [3] M. Brown, W. Grundy, D. Lin, N. Cristianini, C. Sugnet, T. Furey, M. Ares, and D. Haussler, "Knowledge-based analysis of microarray gene expressions data using support vector machines", Tech. Report, University of California in Santa Cruz, 1999. [4] W.S. Cleveland and S.J. Devlin, "Locally Weighted Regression: An Approach to Regression Analysis by Local Fitting", J. Amer. Statist. Assoc. 83, 596-610, 1988 [5] T.M. Cover and P.E. Hart, "Nearest Neighbor Pattern Classification", IEEE Trans. on Information Theory, pp. 21-27, 1967. [6] C. Domeniconi and D. Gunopulos, "Adaptive Nearest Neighbor Classification using Support Vector Machines", Tech. Report UCR-CSE-01-04, Dept. of Computer Science, University of California, Riverside, June 200l. [7] C. Domeniconi, J. Peng, and D. Gunopulos, "An Adaptive Metric Machine for Pattern Classification", Advances in Neural Information Processing Systems, 2000. [8] R.O. Duda and P.E. Hart, Pattern Classification and Scene Analysis. John Wiley & Sons, Inc., 1973. [9] J.H. Friedman "Flexible Metric Nearest Neighbor Classification", Tech. Report, Dept. of Statistics, Stanford University, 1994. [10] T. Hastie and R. Tibshirani, "Discriminant Adaptive Nearest Neighbor Classification", IEEE Trans. on Pattern Analysis and Machine Intelligence, Vol. 18, No.6, pp. 607-615, 1996. [11] T. Joachims, "Text categorization with support vector machines", Pmc. of European Conference on Machine Learning, 1998. [12] T. Joachims, "Making large-scale SVM learning practical" Advances in Kernel Methods - Support Vector Learning, B. Sch6lkopf and C. Burger and A. Smola (ed.), MITPress, 1999. http://www-ai.cs.uni-dortmund.de/thorsten/svm_light.html [13] D.G. Lowe, "Similarity Metric Learning for a Variable-Kernel Classifier", Neural Computation 7(1):72-85, 1995. [14] E. Osuna, R. Freund, and F. Girosi, "Training support vector machines: An application to face detection", Pmc. of Computer Vision and Pattern Recognition, 1997. [15] J.R. Quinlan, C4.5: Programs for Machine Learning. Morgan-Kaufmann Publishers, Inc., 1993. [16] C.J. Stone, Nonparametric regression and its applications (with discussion). Ann. Statist. 5, 595, 1977.
2054 |@word repository:1 proportion:1 duda:1 nd:2 covariance:1 thereby:3 solid:1 initial:1 contains:1 efficacy:1 denoting:1 si:2 assigning:1 john:1 cruz:1 informative:1 girosi:1 remove:1 plot:2 alone:3 greedy:2 intelligence:1 isotropic:1 xk:2 farther:1 provides:2 cse:1 location:2 five:1 adamenn:5 along:13 become:1 consists:1 fitting:1 introduce:1 peng:1 behavior:1 multi:1 chi:2 terminal:1 inspired:1 bellman:1 spherical:1 curse:2 considering:1 increasing:4 becomes:6 provided:2 cleveland:1 bounded:2 moreover:1 maximizes:1 furey:1 burger:1 transformation:4 subclass:2 classifier:8 assoc:1 partitioning:1 unit:2 grant:1 control:1 producing:1 local:17 gunopulos:3 consequence:3 fluctuation:1 might:1 twice:1 relaxing:1 limited:1 sill:3 perpendicular:1 range:1 practical:3 acknowledgment:1 recursive:2 carlotta:2 procedure:1 lla:1 area:3 radial:2 mitpress:1 get:1 doubled:1 undesirable:1 close:12 risk:1 applying:2 influence:2 www:2 elongated:1 demonstrated:2 resolution:4 splitting:2 rule:7 haussler:1 spanned:1 stability:1 construction:1 suppose:1 ilq:3 us:2 element:1 approximated:2 recognition:1 located:1 predicts:1 exk:1 capture:1 worst:2 wj:5 region:2 decrease:1 principled:2 grundy:1 cristianini:1 depend:1 segment:1 serve:1 learner:1 basis:2 separated:3 univ:1 query:17 neighborhood:9 outcome:1 whose:3 xi12:1 quite:1 larger:1 stanford:1 amari:2 enlarges:1 statistic:1 noisy:7 final:1 advantage:2 propose:2 product:2 frequent:1 relevant:2 uci:1 till:1 achieve:1 adapts:1 indi:1 description:1 pll:1 correspondent:1 produce:2 categorization:1 leave:1 liver:2 measured:1 nearest:20 c:3 implies:1 direction:14 modifying:2 quartile:1 centered:1 enable:1 generalization:1 around:4 normal:2 exp:4 mapping:2 predict:3 algorithmic:1 achieves:2 early:1 smallest:1 estimation:1 label:6 sensitive:1 successfully:1 tool:1 weighted:4 minimization:1 clearly:2 gaussian:3 modified:3 ndi:1 ax:1 validated:1 june:1 joachim:2 improvement:5 properly:1 sugnet:1 tech:3 sense:1 dependent:1 nn:3 typically:1 entire:1 iu:1 classification:29 flexible:5 orientation:10 among:1 html:2 spatial:4 equal:2 once:4 extraction:1 having:3 elongation:1 nicely:1 sampling:1 represents:1 argminp:1 report:3 employ:1 randomly:1 dg:1 national:1 dimitrios:1 consisting:2 friedman:2 detection:1 highly:3 severe:1 closer:1 capable:1 respective:1 bq:3 tree:1 euclidean:2 guidance:1 theoretical:1 increased:2 classify:1 soft:1 column:3 cover:1 deviation:6 subset:1 uniform:1 predictor:3 reported:1 constriction:1 corrupted:1 sv:2 traning:1 density:1 stay:1 off:3 squared:3 ek:1 leading:1 potential:1 de:1 summarized:1 inc:2 dann:3 depends:3 crossed:2 later:1 try:1 lowe:1 doing:1 reached:1 bayes:1 relied:1 minimize:2 square:1 variance:3 kaufmann:1 efficiently:1 likewise:1 directional:1 produced:1 dortmund:1 drive:1 ary:1 classified:1 mlearn:1 ed:1 pp:3 conveys:1 associated:2 stop:1 gain:1 proved:1 tunable:1 knowledge:2 dimensionality:2 higher:2 follow:1 reflected:1 arj:2 amer:1 furthermore:2 smola:1 hand:2 effect:3 contain:1 normalized:2 brown:1 vicinity:1 ind:1 ll:1 whereby:1 mlrepository:1 iris:2 stone:1 theoretic:1 demonstrate:1 cp:3 consideration:2 ari:2 common:1 empirically:1 winner:1 exponentially:1 volume:1 jl:1 extend:1 he:5 approximates:1 measurement:1 significant:1 ai:1 stretched:1 tuning:1 dj:1 dot:1 moving:1 similarity:1 surface:1 mlight:2 closest:3 posterior:2 driven:1 scenario:1 arbitrarily:3 yi:2 morgan:1 minimum:1 additional:2 performer:2 employed:1 minl:1 determine:2 maximize:1 ii:1 desirable:1 rj:4 cross:5 offer:2 lin:2 hart:2 award:2 impact:1 prediction:3 emf:1 regression:3 breast:3 vision:1 metric:15 iteration:1 kernel:10 deterioration:1 achieved:3 whereas:2 want:1 median:1 microarray:1 crucial:2 publisher:1 subject:1 tend:1 oq:3 structural:1 presence:1 split:3 variety:2 hastie:1 competing:1 devlin:1 six:1 expression:1 defense:1 svm_light:1 riverside:2 cpt:1 generally:1 santa:1 amount:3 nonparametric:1 dark:1 locally:12 statist:2 svms:10 http:2 exist:1 nsf:2 sign:3 estimated:2 tibshirani:1 vol:1 four:2 procedural:1 drawn:1 sum:1 wu:2 decision:4 ucr:2 scaling:1 bound:1 quadratic:1 constraint:1 scene:1 speed:2 influential:1 department:1 according:2 poor:1 lld:12 em:1 separability:3 wi:1 appealing:4 son:1 osuna:1 making:1 thorsten:1 taken:2 mechanism:2 know:1 junction:1 gaussians:2 available:1 elongate:1 observe:4 eight:2 robustness:1 permanently:1 remaining:1 quinlan:1 uj:1 build:1 seeking:1 move:3 added:1 strategy:3 primary:1 dependence:1 gradient:5 subspace:1 distance:18 constricted:1 simulated:6 outer:1 evenly:1 seven:3 extent:1 discriminant:7 reason:1 ratio:1 pmc:2 difficult:1 mostly:1 pima:3 dci:1 negative:1 guideline:2 adjustable:1 perform:1 sch6lkopf:1 upper:1 vertical:1 observation:4 finite:3 defining:1 variability:1 lfm:18 introduced:1 namely:1 optimized:1 c4:5 california:3 narrow:1 trans:2 address:1 pattern:6 ilxk:2 program:1 including:2 natural:1 rely:1 scheme:9 improve:2 identifies:2 axis:1 carried:2 text:1 determining:1 asymptotic:1 freund:1 expect:1 proportional:1 validation:4 foundation:2 principle:1 pi:3 supported:1 soon:1 offline:1 bias:2 guide:1 neighbor:23 face:1 sparse:1 boundary:27 dimension:4 xn:1 curve:1 scythe:3 computes:3 preventing:1 author:1 commonly:1 adaptive:12 projected:1 bm:3 founded:1 approximate:1 uni:1 gene:1 xi:4 machete:5 sonar:2 table:7 nature:2 robust:2 ca:1 career:1 ignoring:1 improving:1 european:1 spread:1 noise:4 allowed:1 convey:1 augmented:1 join:1 wiley:1 position:1 exponential:4 xl:1 weighting:12 peeling:1 enlarging:1 mitigates:1 minsi:4 decay:1 svm:37 evidence:1 bivariate:1 consist:1 occurring:1 sparseness:1 margin:4 likely:6 infinitely:1 lazy:2 desire:2 applies:2 corresponds:1 satisfies:1 relies:1 conditional:5 goal:2 llxi:1 ann:1 invalid:2 rbf:8 considerable:2 change:3 determined:2 degradation:1 total:1 domeniconi:3 experimental:3 vote:2 formally:1 support:21 relevance:9 dept:3 princeton:1 tested:1 avoiding:1
1,157
2,055
Infinite Mixtures of Gaussian Process Experts Carl Edward Rasmussen and Zoubin Ghahramani Gatsby Computational Neuroscience Unit University College London 17 Queen Square, London WC1N 3AR, England edward,zoubin@gatsby.ucl.ac.uk http://www.gatsby.ucl.ac.uk Abstract We present an extension to the Mixture of Experts (ME) model, where the individual experts are Gaussian Process (GP) regression models. Using an input-dependent adaptation of the Dirichlet Process, we implement a gating network for an infinite number of Experts. Inference in this model may be done efficiently using a Markov Chain relying on Gibbs sampling. The model allows the effective covariance function to vary with the inputs, and may handle large datasets ? thus potentially overcoming two of the biggest hurdles with GP models. Simulations show the viability of this approach. 1 Introduction Gaussian Processes [Williams & Rasmussen, 1996] have proven to be a powerful tool for regression. They combine the flexibility of being able to model arbitrary smooth functions if given enough data, with the simplicity of a Bayesian specification that only requires inference over a small number of readily interpretable hyperparameters ? such as the length scales by which the function varies along different dimensions, the contributions of signal and noise to the variance in the data, etc. However, GPs suffer from two important limitacovariance matrix where is tions. First, because inference requires inversion of an the number of training data points, they are computationally impractical for large datasets. Second, the covariance function is commonly assumed to be stationary, limiting the modeling flexibility. For example, if the noise variance is different in different parts of the input space, or if the function has a discontinuity, a stationary covariance function will not be adequate. Goldberg et al [1998] discussed the case of input dependent noise variance.  Several recent attempts have been aimed at approximate inference in GP models [Williams & Seeger 2001, Smola & Bartlett 2001]. These methods are based on selecting a projection of the covariance matrix onto a smaller subspace (e.g. a subset of the data points) reducing the overall computational complexity. There have also been attempts at deriving more complex covariance functions [Gibbs 1997] although it can be difficult to decide a priori on a covariance function of sufficient complexity which guarantees positive definiteness. In this paper we will simultaneously address both the problem of computational complexity and the deficiencies in covariance functions using a divide and conquer strategy inspired by the Mixture of Experts (ME) architecture [Jacobs et al, 1991]. In this model the input space is (probabilistically) divided by a gating network into regions within which specific separate experts make predictions. Using GP models as experts we gain the double advantage that computation for each expert is cubic only in the number of data point in its region, rather than in the entire dataset, and that each GP-expert may learn different characteristics of the function (such as lengths scales, noise variances, etc). Of course, as in the ME, the learning of the experts and the gating network are intimately coupled. Unfortunately, it may be (practically and statistically) difficult to infer the appropriate number of experts for a particular dataset. In the current paper we sidestep this difficult problem by using an infinite number of experts and employing a gating network related to the Dirichlet Process, to specify a spatially varying Dirichlet Process. An infinite number of experts may also in many cases be more faithful to our prior expectations about complex real-word datasets. Integrating over the posterior distribution for the parameters is carried out using a Markov Chain Monte Carlo approach. Tresp [2001] presented an alternative approach to mixtures of GPs. In his approach both the experts and the gating network were implemented with GPs; the gating network being a softmax of GPs. Our new model avoids several limitations of the previous approach, which are covered in depth in the discussion. 2 Infinite GP mixtures The traditional ME likelihood does not apply when the experts are non-parametric. This is because in a normal ME model the data is assumed to   be  iid  given the  model parameters:             "! #     where and are inputs and outputs (boldface denotes vectors), are the parameters of expert  , ! are the parameters of the gating network and  are the discrete indicator variables assigning data points to experts. This iid assumption is contrary to GP models which solely model the dependencies in the joint distribution (given the hyperparameters). There is a joint distribution corresponding to every possible assignment of data points to experts; therefore the likelihood is a sum over (exponentially many) assignments:   $  %&' ()  * +,(-  "! 65   5  6    &/.  01324  )78924  )7* ;:<,(-  "! #= (1) Given the configuration (>?,@ 4=A=A=4 $<BC , the distribution factors into the product, over experts, of the joint Gaussian distribution of all data points assigned to each expert. Whereas the original ME formulation used expectations of assignment variables called responsibili ties, this is inadequate for inference in the mixture of GP experts. Consequently, we directly  represent the indicators, , and Gibbs sample for them to capture their dependencies. In Gibbs sampling we need the posterior conditional distribution for each indicator given  indicators     all the remaining and the data:   , D (FE $ G $ C "! H%  IJ "(FE $   D (FE $ "! # where (CE denotes all indicators except number K . We defer discussion of the second term defining the gating network to the next section. As discussed, the first term being the likelihood given the indicators factors into independent terms for each expert. For Gibbs 5MON sampling we therefore need the probability of output  5 under GP number  :  924L  K $PLC7* <2QCL PL)7 #= For a GP model, this conditional density is the well known Gaussian [Williams & Rasmussen, 1996]:   E @ E       6       $      E @         E  $      # (2) where the covariance matrix depends on the parameters . Thus, for the GP expert, we compute the above conditional density by simply evaluating the GP on the data assigned to it. Although this equation looks computationally expensive, we can keep track of the inverse covariance matrices and reuse them for consecutive Gibbs updates by performing rank one updates (since Gibbs sampling changes at most one indicator at a time). We are free to choose any valid covariance function for the experts. In our simulations we employed the following Gaussian covariance function:   $ &  $ $+-,   G  "#! %$I 0   ('()*  @ .  K K0/ (3) 1 controlling the signal variance,  @ controlling the noise variance, $ with hyperparameters ) and controlling the length scale or (inverse) relevance of the 2 -th dimension of  in relation to predicting  ; . is the Kronecker delta function (i.e. .  K K /  ! if K  K / , o.w. 0). 3 The Gating network The gating network assigns probability to different experts based entirely on the input. We will derive a gating network based on the Dirichlet Process which can be defined as the limit of a Dirichlet distribution when the number of classes tends to infinity. The standard Dirichlet Process is not input dependent, but we will modify it to serve as a gating mechanism. We start from a symmetric Dirichlet distribution on proportions: 43 @ A= = = 365C 76 - 8:9&;9=<?>A@&BQC7 '%D G  E 4 76 5  G3HJI 5 E @ E 4 7 F' D  2000] that where 7 is the (positive) concentration parameter. It can be shown [Rasmussen, 3 the conditional probability of a single indicator when integrating over the variables and K  letting D tend to infinity is given by: E components where all other components 4K  combined: &0R   K MLGN  1       ( E  :  N &  N    for all K /  K  (CE 1   76  76  6E , O ! 7 7 , O ! 7 (4) where GE ( QP S . , 8 ) is the occupation number of expert  excluding observation K , and is the total number of data points. This shows that the probabilities are proportional to the occupation numbers. To make the gating network input dependent, we will simply employ a local estimate 1 for this occupation classifier:  a kernel    R number JUWV  using E 4K    T ! P S   R F W U  V   . ,&  8 P S   U (5) where the delta function selects data points assigned to class  , and is the kernel function parametrized by ! . As an example we use a Gaussian kernel function: UOV 1    $   $ $+   G X1  Y  #!  $    ' !  (6) this local estimate won?t generally be an integer, but this doesn?t have any adverse consequences $ parameterized by length scales ! for each dimension. These length scales allow dimensions of  space to be more or less relevant to the gating network classification. We Gibbs sample from the indicator variables by multiplying the input-dependent Dirichlet process prior eq. (4) and (5) with the GP conditional density eq. (2). Gibbs sampling in an infinite model requires that the indicator variables can take on values that no other indicator variable has already taken, thereby creating new experts. We use the auxiliary variable approach of Neal [1998] (algorithm 8 in that paper). In this approach hyperparameters for new experts are sampled from their prior and the likelihood is evaluated based on these. This requires finding the likelihood of a Gaussian process with no data. Fortunately, for, the covariance function eq. (3) this likelihood is Gaussian with zero mean and variance   @ . If all data points are assigned to a single GP, the likelihood calculation will still be cubic in the number of data points (per Gibbs sweep over all indicators). We can reduce the computational complexity by introducing the constraint that no GP expert can have more than 2 max data points assigned to it. This is easily implemented by modifying the conditionals in the Gibbs sampler. The hyperparameter 7 controls the prior probability of assigning a data point to a new expert, and therefore influences the total number of experts used to model the data. As in Rasmussen [2000], we give a vague inverse gamma prior to 7 , and sample from its posterior using Adaptive Rejection Sampling (ARS) [Gilks & Wild, 1992]. Allowing 7 to vary gives the model more freedom to infer the number of GPs to use for a particular dataset. Finally we need to do inference for the parameters of the gating function. Given a set of indicator variables one could use standard methods from kernel classification to optimize the kernel widths in different directions. These methods typically optimize the leave-oneout pseudo-likelihood (ie the product of the conditionals), since computing the likelihood in a model defined purely from conditional distributions as in eq. (4), (5) & (6) is generally difficult (and as pointed out in the discussion section there may not even be a single likelihood). In our model we multiply the pseudo-likelihood by a (vague) prior and sample from the resulting pseudo-posterior. 4 The Algorithm The individual GP experts are given a stationary Gaussian covariance function,, with a sin# (where gle length scale per dimension, a signal variance and a noise variance, i.e. is the dimension of the input) hyperparameters per expert, eq. (3). The signal and noise variances are given inverse gamma priors with hyper-hypers and (separately for the two variances). This serves to couple the hyperparameters between experts, and allows the priors on   and *@ (which are used when evaluating auxiliary classes) to adapt. Finally we give vague independent log normal priors to the lenght scale paramters ) and ! .   The algorithm for learning an infinite mixture of GP experts consists of the following steps: 1. 2. 3. 4. 5. 2 Initialize indicator variables   to a single value (or a few values if individual GPs are to be kept small for computational reasons). Do a Gibbs sampling sweep over all indicators. $ Do Hybrid Monte Carlo (HMC) [Duane et al, 1987] for hyperparameters of the     @ ) GP covariance function, , for each expert in turn. We used 10 leapfrog iterations with a stepsize small enough that rejections were rare. Optimize the hyper-hypers, & , for each of the variance parameters. Sample the Dirichlet process concentration parameter, 7 using ARS.   We simply set the conditional probability of joining a class which has been deemed full to zero. 100 50 50 Acceleration (g) Acceleration (g) 100 0 ?50 ?50 ?100 ?100 iMGPE stationary GP ?150 0 0 10 20 30 40 Time (ms) 50 60 ?150 0 10 20 30 40 Time (ms) 50 60 Figure 1: The left hand plot shows the motorcycle impact data (133 points) together with the median of the model?s predictive distribution, and for comparison the mean of a staNJN tionary covariance GP model (with optimized hyperparameters). On the right hand plot we show ! samples from the posterior distribution for the iMGPE of the (noise free) function evaluated intervals of 1 N ms. We have jittered the points in the plot along the time = # ms noise, so that the density can be seen more easily. dimension by adding uniform # std error ( ) confidence interval for the (noise free) function predicted by Also, the a stationary GP is plotted (thin lines). 6. Sample the gating kernel widths, ! ; we use the Metropolis method to sample from the pseudo-posterior with a Gaussian proposal fit at the current ! 3 7. Repeat from 2 until the Markov chain has adequately sampled the posterior. 5 Simulations on a simple real-world data set To illustrate our algorithm, we used the motorcycle dataset, fig. 1, discussed in Silverman [1985]. This dataset is obviously non-stationary and has# input-dependent noise. We noticed that the raw data is discretized# into bins of size  ! = g; accordingly we cut off the prior for the noise variance at   ' ! . The model is able to capture the general shape of the function and also the input-dependent N the right hand plot in fig. 1, where the nature of the noise (fig. 1). This can be seen from uncertainty of the function is very low for !  owing to a small inferred noise level in this region. For comparison, the predictions from a stationary GP has been superimposed in fig. 1. Whereas the medians of the predictive distributions agree to a large extent (left N (right hand). The hohand plot), we see a huge difference in the predictive distributions  L  J N moscedastic GP cannot capture the very tight distribution for  ! ms offered by iMGPE. Also for large ms, the iMGPE model predicts with fairly high probability that the signal could be very close to zero. Note that the predictive distribution of the function is multimodal, for example, around time 35 ms. Multimodal predictive distributions could in principle be obtained from an ordinary GP by integrating over hyperparameters, howN ever, in a mixture of GP?s model they can arise naturally. The predictive distribution of the function appears to have significant mass around g which seems somewhat artifactual. We explicitly did not normalize or center the data, which has a large range in output. The 3 The Gaussian fit uses the derivative and Hessian of the log posterior wrt the log length scales. Since this is an asymmetric proposal the acceptance probabilities must be modified accordingly. This scheme has the advantage of containing no tunable parameters; however when the dimension  is large, it may be computationally more efficient to use HMC, to avoid calculation of the Hessian. 14 100 12 90 20 80 70 60 60 50 80 10 frequency 40 8 6 40 4 30 100 20 120 10 0 20 40 60 80 100 120 2 0 5 10 15 20 25 number of occupied experts 30 Figure 2: The left hand plot shows the number of times, out of 100 samples, that the indicator variables for two data points were equal. The data have been sorted from left-toright according to the value of the time variable (since the data is not equally spaced in time the axis of this matrix cannot be aligned with the plot in fig.1). The right hand plot shows a histogram over the 100 samples of the number of GP experts used to model the data. Gaussian processes had zero mean a priori, which coupled with the concentration of data around zero may explain the posterior mass at zero. It would be more natural to treat the GP means as separate hyperparameters controlled by a hyper-hyperparameter (centered at zero) and do inference on them, rather than fix them all at 0. Although the scatter of data from the predictive distribution for iMGPE looks somewhat messy with multimodality etc, it is important to note that it assigns high density to regions that seem probable. The motorcycle data appears to have roughly three regions: a flat low-noise region, followed by a curved region, and a flat high noise region. This intuition is bourne out by the model. We can see this in two ways. Fig 2. (left) shows the number of times two data points were assigned to the same expert. A N clearly defined block captures the initial flat region and a few other points that lie near the g line; the middle block captures the curved region, with a more gradual transition to the last flat region. A histogram of the number of GP experts used showsN that the posterior distribution of number of needed GPs has a broadN peak between and ! , where less than 3 occupied experts is very unlikely, and above ! becoming progressively less likely. Note that it never uses just a single GP to model the data which accords with the intuition that a single stationary covariance function would be inadequate. We should point out that the model is not trying to do model selection between finite GP mixtures, but rather always assumes that there are infinitely many available, most of which contribute with small mass 4 to a diffuse density in the background. In figure 3 we assessed the convergence rate of the Markov Chain by plotting the autoNJN%N mixing time is around correlation function for several parameters. We conclude that the J N % N % N N 5N 100 iterations . Consequently, we run the chain for a total of !J! iterations, discarding the initial ! (burn-in) and keeping every ! ?th after that. The total computation time was around 1 hour (1 GHz Pentium). The right hand panel of figure 3 shows the distribution of the gating function kernel width The total mass of the non-represented experts is    , where the posterior for  in this experiment is peaked between and (see figure 3, bottom right panel), corresponding to about  of the total mass 5 the sum of the auto-correlation coefficients from  to  is an estimate of the mixing time 4 0.8 5 0.6 0 0.4 10 frequency auto correlation coefficient log number of occupied experts log gating kernel width log Dirichlet concentration frequency 10 1 0.2 0 0 50 100 150 time lag in iterations 200 ?1 ?0.5 0 0.5 log (base 10) gating function kernel width 5 0 ?0.5 0 0.5 1 log (base 10) Dirichlet process concentration NJN%left N hand plot shows the auto-correlation for various parameters of the model Figure 3: The N%N of the (log) kernel based on !%! iterations. The right hand plots show the distribution width ! and (log) Dirichlet concentration parameter 7 , based on ! samples from the posterior. N and the concentration parameter of the Dirichlet process. The posterior 6 kernel width ! = lies between ! and ; comparing to the scale of the inputs these are quite short distances, corresponding to rapid transitions between experts (as opposed to lengthy intervals with multiple active experts). This corresponds well with our visual impression of the data. ! 6 Discussion and Conclusions We now return to Tresp [2000]. There are four ways in which the infinite mixture of GP experts differs from, and we believe, improves upon the model presented by Tresp. First, in his model, although a gating network divides up the input space, each GP expert predicts on the basis of all of the data. Data that was not assigned to a GP expert can therefore spill over into predictions of a GP, which will lead to bias near region boundaries especially for experts with long length scales. Second, if there are experts, Tresp?s model has GPs (the experts, noise models, and separate gating functions) each of which requires 4 computations. In our model computations over the entire dataset resulting in  since the experts divide up the data points, if there are experts equally dividing the data computations (each of Gibbs updates requires a rank-one an iteration takes   '  for each of the experts andN the Hybrid Monte Carlo takes computation   '     ' times ). Even for modest (e.g. ! ) this can be a significant saving. Inference for the gating length scale parameters is    if the full Hessian is used, but can be reduced to   for a diagonal approximation, or Hybrid Monte Carlo if the input dimension is large. Third, by going to the Dirichlet process infinite limit, we allow the model to infer the number of components required to capture the data. Finally, in our model the GP hyperparameters are not fixed but are instead inferred from the data. We have defined the gating network prior implicitly in terms of the conditional distribution of an indicator variable given all the other indicator variables. Specifically, the distribution of this indicator variable is an input-dependent Dirichlet process with counts given by local estimates of the data density in each class eq. (5). We have not been able to prove that these conditional distributions are always consistent with a single joint distribution over for comparison the (vague) prior on the kernel width is log normal with   of the mass between   and , corresponding to very short (sub sample) distances upto distances comparable to the entire input range  6 the indicators. If indeed there does not exist a single consistent joint distribution the Gibbs sampler may converge to different distributions depending on the order of sampling. We are encouraged by the preliminary results obtained on the motorcycle data. Future work should also include empirical comparisons with other state-of-the-art regression methods on multidimensional benchmark datasets. We have argued here that single iterations of the MCMC inference are computationally tractable even for large data sets, experiments will show whether mixing is sufficiently rapid to allow practical application. We hope that the extra flexibility of the effective covariance function will turn out to improve performance. Also, the automatic choice of the number of experts may make the model advantageous for practical modeling tasks. Finally, we wish to come back to the modeling philosophy which underlies this paper. The computational problem in doing inference and prediction using Gaussian Processes arises out of the unrealistic assumption that a single covariance function captures the behavior of the data over its entire range. This leads to a cumbersome matrix inversion over the entire data set. Instead we find that by making a more realistic assumption, that the data can be modeled by an infinite mixture of local Gaussian processes, the computational problem also decomposes into smaller matrix inversions. References Gibbs, M. N. (1997). Bayesian Gaussian Processes for Regression and Classification. PhD thesis. University of Cambridge. Goldberg, P. W., Williams, C. K. I., & Bishop C. M. (1998). Regression with Inputdependent Noise, NIPS 10. Duane, S., Kennedy, A. D., Pendleton, B. J., and Roweth, D. (1987). Hybrid Monte Carlo, Physics letters B, vol. 55, pp. 2774?2777. Gilks, W. R. & Wild, P. (1992). Adaptive rejection sampling for Gibbs sampling. Applied Statistics 41, 337?348. Jacobs, R. A., Jordan, M. I., Nowlan, S. J. & Hinton, G. E. (1991). Adaptive mixture of local experts. Neural Computation, vol 3, pp 79?87. Neal, R. M. (1998). Markov chain sampling methods for Dirichlet process mixture models. Technical Report 4915, Department of Statistics, University of Toronto. http://www.cs.toronto.edu/ radford/mixmc.abstract.html. Rasmussen, C. E. (2000). The Infinite Gaussian Mixture Model, NIPS 12, S.A. Solla, T.K. Leen and K.-R. M?uller (eds.), pp. 554?560, MIT Press. Silverman, B. W. (1985). Some aspects of the spline smoothing approach to non-parametric regression curve fitting. J. Royal Stat. Society. Ser. B, vol. 47, pp. 1?52. Smola A. J. and Bartlett, P. (2001). Sparse Greedy Gaussian Process Regression, NIPS 13. Tresp V. (2001). Mixtures of Gaussian Process, NIPS 13. Williams, C. K. I. and Seeger, M. (2001). Using the Nystr?om Method to Speed Up Kernel Machines, NIPS 13. Williams, C. K. I. and C. E. Rasmussen (1996). Gaussian Processes for Regression, in D. S. Touretzky, M. C. Mozer and M. E. Hasselmo (editors), NIPS 8, MIT Press.
2055 |@word middle:1 inversion:3 advantageous:1 proportion:1 seems:1 simulation:3 gradual:1 covariance:18 jacob:2 thereby:1 nystr:1 initial:2 configuration:1 selecting:1 bc:1 current:2 comparing:1 nowlan:1 assigning:2 scatter:1 must:1 readily:1 realistic:1 shape:1 plot:10 interpretable:1 update:3 progressively:1 stationary:8 greedy:1 accordingly:2 short:2 contribute:1 toronto:2 along:2 consists:1 prove:1 combine:1 wild:2 fitting:1 multimodality:1 indeed:1 rapid:2 behavior:1 roughly:1 discretized:1 inspired:1 relying:1 panel:2 mass:6 finding:1 impractical:1 guarantee:1 pseudo:4 every:2 multidimensional:1 tie:1 classifier:1 uk:2 control:1 unit:1 ser:1 positive:2 local:5 modify:1 tends:1 limit:2 consequence:1 treat:1 joining:1 solely:1 becoming:1 burn:1 range:3 statistically:1 faithful:1 gilks:2 practical:2 block:2 implement:1 differs:1 silverman:2 empirical:1 projection:1 word:1 integrating:3 confidence:1 zoubin:2 onto:1 cannot:2 close:1 selection:1 influence:1 www:2 optimize:3 center:1 williams:6 simplicity:1 assigns:2 deriving:1 his:2 handle:1 limiting:1 controlling:3 carl:1 gps:8 goldberg:2 us:2 expensive:1 std:1 asymmetric:1 cut:1 predicts:2 bottom:1 capture:7 region:12 solla:1 intuition:2 mozer:1 complexity:4 messy:1 tight:1 predictive:7 serve:1 purely:1 upon:1 basis:1 vague:4 easily:2 joint:5 multimodal:2 k0:1 represented:1 various:1 effective:2 london:2 monte:5 hyper:3 pendleton:1 mon:1 lag:1 quite:1 statistic:2 gp:35 obviously:1 advantage:2 ucl:2 product:2 adaptation:1 relevant:1 motorcycle:4 aligned:1 mixing:3 flexibility:3 normalize:1 convergence:1 double:1 leave:1 tions:1 derive:1 illustrate:1 ac:2 stat:1 depending:1 eq:6 edward:2 dividing:1 implemented:2 auxiliary:2 predicted:1 come:1 c:1 direction:1 owing:1 modifying:1 centered:1 bin:1 argued:1 fix:1 preliminary:1 probable:1 extension:1 pl:2 practically:1 around:5 sufficiently:1 normal:3 vary:2 consecutive:1 hasselmo:1 tool:1 hope:1 uller:1 mit:2 clearly:1 gaussian:20 always:2 modified:1 rather:3 occupied:3 avoid:1 varying:1 probabilistically:1 leapfrog:1 rank:2 likelihood:11 superimposed:1 seeger:2 pentium:1 inference:10 dependent:8 entire:5 typically:1 unlikely:1 relation:1 going:1 selects:1 overall:1 classification:3 html:1 priori:2 art:1 softmax:1 initialize:1 fairly:1 smoothing:1 equal:1 never:1 saving:1 sampling:11 encouraged:1 look:2 thin:1 bourne:1 peaked:1 future:1 report:1 spline:1 employ:1 few:2 simultaneously:1 gamma:2 individual:3 attempt:2 freedom:1 huge:1 acceptance:1 multiply:1 mixture:15 wc1n:1 chain:6 bq:1 modest:1 divide:3 plotted:1 roweth:1 modeling:3 ar:3 queen:1 assignment:3 ordinary:1 introducing:1 subset:1 rare:1 uniform:1 inadequate:2 dependency:2 varies:1 jittered:1 combined:1 density:7 peak:1 ie:1 off:1 physic:1 together:1 thesis:1 containing:1 choose:1 opposed:1 creating:1 expert:56 sidestep:1 derivative:1 return:1 coefficient:2 explicitly:1 depends:1 doing:1 start:1 defer:1 contribution:1 om:1 square:1 variance:13 characteristic:1 efficiently:1 spaced:1 inputdependent:1 bayesian:2 raw:1 iid:2 carlo:5 multiplying:1 kennedy:1 explain:1 cumbersome:1 touretzky:1 ed:1 lengthy:1 c7:2 frequency:3 pp:4 naturally:1 couple:1 gain:1 sampled:2 dataset:6 tunable:1 improves:1 back:1 appears:2 specify:1 formulation:1 done:1 evaluated:2 leen:1 just:1 smola:2 until:1 correlation:4 hand:9 believe:1 adequately:1 assigned:7 spatially:1 symmetric:1 neal:2 sin:1 width:8 won:1 m:7 trying:1 impression:1 qp:1 exponentially:1 discussed:3 significant:2 cambridge:1 gibbs:16 automatic:1 pointed:1 had:1 specification:1 etc:3 base:2 posterior:13 recent:1 seen:2 fortunately:1 somewhat:2 employed:1 converge:1 signal:5 full:2 multiple:1 infer:3 smooth:1 technical:1 england:1 calculation:2 adapt:1 long:1 divided:1 equally:2 controlled:1 impact:1 prediction:4 underlies:1 regression:8 expectation:2 iteration:7 represent:1 kernel:13 histogram:2 accord:1 proposal:2 whereas:2 hurdle:1 conditionals:2 separately:1 interval:3 background:1 median:2 extra:1 tend:1 contrary:1 seem:1 jordan:1 integer:1 near:2 viability:1 enough:2 fit:2 architecture:1 reduce:1 whether:1 bartlett:2 reuse:1 suffer:1 hessian:3 adequate:1 generally:2 covered:1 aimed:1 gle:1 reduced:1 http:2 exist:1 neuroscience:1 delta:2 track:1 per:3 hji:1 paramters:1 discrete:1 hyperparameter:2 vol:3 four:1 ce:2 kept:1 sum:2 run:1 inverse:4 parameterized:1 powerful:1 uncertainty:1 letter:1 decide:1 toright:1 comparable:1 entirely:1 followed:1 kronecker:1 deficiency:1 infinity:2 constraint:1 flat:4 diffuse:1 aspect:1 speed:1 performing:1 department:1 according:1 smaller:2 intimately:1 g3:1 metropolis:1 making:1 taken:1 computationally:4 equation:1 agree:1 turn:2 count:1 mechanism:1 wrt:1 needed:1 letting:1 ge:1 tractable:1 serf:1 available:1 apply:1 appropriate:1 upto:1 stepsize:1 alternative:1 original:1 denotes:2 dirichlet:16 remaining:1 assumes:1 include:1 spill:1 ghahramani:1 especially:1 conquer:1 society:1 sweep:2 noticed:1 already:1 strategy:1 parametric:2 concentration:7 traditional:1 diagonal:1 subspace:1 distance:3 separate:3 parametrized:1 me:6 extent:1 reason:1 boldface:1 length:9 modeled:1 difficult:4 unfortunately:1 hmc:2 fe:3 potentially:1 oneout:1 allowing:1 observation:1 markov:5 datasets:4 benchmark:1 finite:1 curved:2 defining:1 hinton:1 excluding:1 ever:1 arbitrary:1 overcoming:1 inferred:2 required:1 optimized:1 hour:1 discontinuity:1 nip:6 address:1 able:3 max:1 royal:1 unrealistic:1 natural:1 hybrid:4 predicting:1 indicator:20 scheme:1 improve:1 axis:1 carried:1 deemed:1 coupled:2 auto:3 tresp:5 prior:12 occupation:3 limitation:1 proportional:1 proven:1 offered:1 sufficient:1 consistent:2 principle:1 plotting:1 editor:1 course:1 repeat:1 last:1 rasmussen:7 free:3 keeping:1 bias:1 allow:3 sparse:1 ghz:1 curve:1 boundary:1 dimension:9 depth:1 evaluating:2 avoids:1 valid:1 doesn:1 world:1 transition:2 commonly:1 adaptive:3 employing:1 approximate:1 implicitly:1 keep:1 active:1 assumed:2 conclude:1 decomposes:1 hown:1 learn:1 nature:1 complex:2 cl:1 did:1 noise:18 hyperparameters:11 arise:1 fig:6 biggest:1 cubic:2 gatsby:3 definiteness:1 sub:1 wish:1 lie:2 third:1 specific:1 discarding:1 bishop:1 gating:23 adding:1 phd:1 tionary:1 rejection:3 artifactual:1 simply:3 likely:1 infinitely:1 visual:1 radford:1 duane:2 corresponds:1 conditional:9 sorted:1 consequently:2 acceleration:2 change:1 adverse:1 infinite:11 except:1 reducing:1 specifically:1 sampler:2 called:1 total:6 andn:1 college:1 arises:1 assessed:1 relevance:1 philosophy:1 mcmc:1
1,158
2,056
Associative memory in realistic neuronal networks P.E. Latham* Department of Neurobiology University of California at Los Angeles Los Angeles, CA 90095 pel@ucla.edu Abstract Almost two decades ago , Hopfield [1] showed that networks of highly reduced model neurons can exhibit multiple attracting fixed points, thus providing a substrate for associative memory. It is still not clear, however, whether realistic neuronal networks can support multiple attractors. The main difficulty is that neuronal networks in vivo exhibit a stable background state at low firing rate, typically a few Hz. Embedding attractor is easy; doing so without destabilizing the background is not. Previous work [2, 3] focused on the sparse coding limit, in which a vanishingly small number of neurons are involved in any memory. Here we investigate the case in which the number of neurons involved in a memory scales with the number of neurons in the network. In contrast to the sparse coding limit, we find that multiple attractors can co-exist robustly with a stable background state. Mean field theory is used to understand how the behavior of the network scales with its parameters, and simulations with analog neurons are presented. One of the most important features of the nervous system is its ability to perform associative memory. It is generally believed that associative memory is implemented using attractor networks - experimental studies point in that direction [4- 7], and there are virtually no competing theoretical models. Perhaps surprisingly, however, it is still an open theoretical question whether attractors can exist in realistic neuronal networks. The "realistic" feature that is probably hardest to capture is the steady firing at low rates - the background state - that is observed throughout the intact nervous system [8- 13]. The reason it is difficult to build an attractor network that is stable at low firing rates, at least in the sparse coding limit, is as follows [2,3]: Attractor networks are constructed by strengthening recurrent connections among sub-populations of neurons. The strengthening must be large enough that neurons within a sub-population can sustain a high firing rate state, but not so large that the sub-population can be spontaneously active. This implies that the neuronal gain functions - the firing rate of the post-synaptic neurons as a function of the average ? http) / culture.neurobio.ucla.edu/ "'pel firing rate of the pre-synaptic neurons - must be sigmoidal: small at low firing rate to provide stability, high at intermediate firing rate to provide a threshold (at an unstable equilibrium), and low again at high firing rate to provide saturation and a stable attractor. In other words, a requirement for the co-existence of a stable background state and multiple attractors is that the gain function of the excitatory neurons be super linear at the observed background rates of a few Hz [2,3]. However - and this is where the problem lies - above a few Hz most realistic gain function are nearly linear or sublinear (see, for example, Fig. Bl of [14]). The superlinearity requirement rests on the implicit assumption that the activity of the sub-population involved in a memory does not affect the other neurons in the network. While this assumption is valid in the sparse coding limit , it breaks down in realistic networks containing both excitatory and inhibitory neurons. In such networks, activity among excitatory cells results in inhibitory feedback. This feedback, if powerful enough, can stabilize attractors even without a saturating nonlinearity, essentially by stabilizing the equilibrium (above considered unstable) on the steep part of the gain function. The price one pays, though, is that a reasonable fraction of the neurons must be involved in each of the memories, which takes us away from the sparse coding limit and thus reduces network capacity [15]. 1 The model A relatively good description of neuronal networks is provided by synaptically coupled, conductance-based neurons. However, because communication is via action potentials, such networks are difficult to analyze. An alternative is to model neurons by their firing rates. While this is unlikely to capture the full temporal network dynamics [16], it is useful for studying equilibria. In such simplified models, the equilibrium firing rate of a neuron is a function of the firing rates of all the other neurons in the network. Letting VEi and VIi denote the firing rates of the excitatory and inhibitory neurons, respectively, and assuming that synaptic input sums linearly, the equilibrium equations may be written ?Ei (~Af;EVEj' ~Af;'V'j) (la) ?;; (~AifVEj, ~ Ai!V,j) . (lb) Here ?E and ?I are the excitatory and inhibitory gain functions and Aij determines the connection strength from neuron j to neuron i. The gain functions can, in principle, be derived from conductance-based model equations [17]. Our goal here is to determine under what conditions Eq. (1) allows both attractors and a stable state at low firing rate. To accomplish this we will use mean field theory. While this theory could be applied to the full set of equations, to reduce complexity we make a number of simplifications. First, we let the inhibitory neurons be completely homogeneous (?Ii independent of i and connectivity to and from inhibitory neurons all-to-all and uniform). In that case, Eq. (lb) becomes simply VI = ?(VE' VI) where VE and VI are the average firing rates of the excitatory and inhibitory neurons. Solving for VI and inserting the resulting expression into Eq. (la) results in the expression VEi = ?Ei(LjAijEVEj,AEIVI(VE)) where A EI == LjAijI. Second, we let cP Ei have the form cP Ei (u, v) = cP E( Xi + bu - ev) where Xi is a Gaussian random variable, and similarly for cPT (except with different constants band e and no dependence on i). Finally, we assume that cPT is threshold linear and the network operates in a regime in which the inhibitory firing rate is above zero. With these simplifications, and a trivial redefinition of constants, Eq. (la) becomes (2) We have dropped the sub and superscript E, since Eq. (2) refers exclusively to excitatory neurons, defined v to be the average firing rate, v == N-1 Li Vi, and rescaled parameters. We let the function cP be 0(1), so f3 can be interpreted as the gain. The parameter p is the number of memories. The reduction from Eq. (1) to Eq. (2) was done solely to simplify the analysis; the techniques we will use apply equally well to the general case, Eq. (1). Note that the gain function in Eq. (2) decreases with increasing average firing rate, since it's argument is -(1 + a)v and a is positive. This negative dependence on v arises because we are working in the large coupling regime in which excitation and inhibition are balanced [18,19]. The negative coupling to firing rate has important consequences for stability, as we will see below. We let the connectivity matrix have the form Here N is the number of excitatory neurons; Cij , which regulates the degree of connectivity, is lie with probability e and and 0 with probability (1 - e) (except Cii = 0, meaning no autapses); g(z) is an 0(1) clipping function that keeps weights from falling below zero or getting too large; (g) is the mean value of g(z), defined in Eq. (4) below; W i j , which corresponds to background connectivity, is a random matrix whose elements are Gaussian distributed with mean 1 and variance 8w 2 ; and J ij produces the attractors. We will follow the Hopfield prescription and write J ij as (3) where f is the coupling strength among neurons involved in the memories, and the patterns TJ",i determine which neurons participate in each memory. The TJ",i are a set of uncorrelated vectors with zero mean and unit variance. In simulations we use TJ",i = [(1 - 1)11]1/2 with probability 1 and -(f 1(1 - IW /2 with probability 1 - I, so a fraction 1 of the neurons are involved in each memory. Other choices are unlikely to significantly change our results. 2 Mean field equations The main difficulty in deriving the mean field equations from Eq. (2) is separating the signal from the noise. Our first step in this endeavor is to analyze the noise associated with the clipped weights. To do this we break Cijg(Wij pieces: Cijg(Wij + Jij) = (g) + (g')Jij + bCij where + J ij ) into two The angle brackets around 9 represent an average over the distributions of W ij and Jij, and a prime denotes a derivative. In the large p limit, bCij can be treated as a random matrix whose main role is to increase the effective noise [20]. The mean of bCij is zero and its variance normalized to (g)2 / c, which we denote (Y2, is given by For large p, the elements of Jij are Gaussian with zero mean and variance the averages involving 9 can be written E2, so (4) where k can be either an exponent or a prime and the "I" in g(1 + z) corresponds to the mean of W ij . In our simulations we use the clipping function g(z) = z if z is between 0 and 2, 0 if z ::::; 0 and 2 if z ;::: 2. Our main assumptions in the development of a mean field theory are that L;#i bCijvj is a Gaussian random variable, and that bCij and Vj are independent. Consequently, where (v 2 ) == N- 1 L;i v; is the second moment of the firing rate. Letting 8i be a zero mean Gaussian random variable with variance 82 == (Y2 (v 2)/ cN, we can use the above assumptions along with the definition of Jij , Eq. (3), to write Eq. (20) as (5) We have defined the clipped memory strength, Ee , as Ee == E(g')/(g). While it is not totally obvious from the above equations, it can be shown that both (Y2 and Ee become independent of E for large E. This makes network behavior robust to changes in E, the strength of the memories, so long as E is large. Derivation ofthe mean field equations from Eq. (5) follow standard methods [21,22]. For definiteness we take ?(x) to be threshold linear: ?(x) = max(O, x). For the case of one active memory, the mean field equations may then be written in the form {3Ec w + q ) (6a) 1] (32E~ [1J 2 a(l-r)2 CE~+(1-q)2 [F2(z)+jflF2(w ,z)] 1 r ( 1- r flF1 w,z (6b) {32B 2a2/x2 (1 ~ r)2 a [Fl (z) + j flFl (w, zW a{3Ecq 1-q (3E~ 1+a (6c) [Fo(z) + jflFo(w,z)] (6d) Ec where a == piN is the load parameter, Xo and B6/P are the mean and variance of of Xi (see Eq. (2)), and, recall, j is the fraction of neurons that participate in each memory. The functions Fk and flFk are defined by 1 00 -z Fdw d~ k 2 (27r )1/2 (z +~) exp( -~ /2) + z) - Fk( Z) . For large negative z, Fk(z) vanishes as exp(-z2/2) , while for large positive z, Fk(Z) --+ zk /k!. The average firing rate, v, and strength of the memory, m == N- 1 2:: i rJljVj (taken without loss of generality to be the overlap with pattern 1), are given in terms of z and was v Xo m 3 Results The mean field equations can be understood by examining Eqs. (6a) and (6b). The first of these, Eq. (6a), is a rescaled form of the equation for the overlap, m. (From the definition of flFt given above, it can be seen that m is proportional to w for small w). This equation always has a solution at w = 0 (and thus m = 0) , which corresponds to a background state with no memories active. If {3Ec is large enough, there is a second solution with w (and thus m) greater than zero. This second solution corresponds to a memory. The other relevant equation, Eq. (6b), describes the behavior of the mean firing rate. This equation looks complicated only because the noise - the variation in firing rate from neuron to neuron - must be determined self-consistently. The solutions to Eqs. (6a) and (6b) are plotted in Fig. 1 in the z-w plane. The solid lines, including the horizontal line at w = 0, represents the solution to Eq. (6a), the w , ~ ',.: ... t t ... w=o z Figure 1: Graphical solution of Eqs. (6a) and (6b). Solid lines, including the one at w = 0: solution to Eq. (6a). Dashed line: solution to Eq. (6b). The arrows indicate approximate flow directions: vertical arrows indicate time evolution of w at fixed z; horizontal arrows indicate time evolution of z at fixed w. The black squares show potentially stable fixed points. Note the exchange of stability to the right of the solid curve, indicating that intersections too far to the right will be unstable. dashed line the solution to Eq. (6b), and their intersections solutions to both. While stability cannot be inferred from the equilibrium equations, a reasonable assumption is that the evolution equations for the firing rates , at least near an equilibrium, have the form Tdvi/dt = ?i - Vi. In that case, the arrows represent flow directions, and we see that there are potentially stable equilibria at the intersections marked by the solid squares. Note that in the sparse coding limit, f ---+ 0, z is independent of w, meaning that the mean firing rate, v , is independent of the overlap, m. In this limit there can be no feedback to inhibitory neurons , and thus no chance for stabilization. In terms of Fig. 1, the effect of letting f ---+ 0 is to make the dashed line vertical. This eliminates the possibility of the upper stable equilibrium (the solid square at w > 0), and returns us to the situation where a superlinear gain function is required for attractors to be embedded, as discussed in the introduction. Two important conclusions can be drawn from Fig. 1. First, the attractors can be stable even though the gain functions never saturate (recall that we used thresholdlinear gain functions). The stabilization mechanism is feedback to inhibitory neurons, via the -(1 + a)v term in Eq. (2). This feedback is what makes the dashed line in Fig. 1 bend, allowing a stable equilibrium at w > O. Second, if the dashed line shifts to the right relative to the solid line, the background becomes destabilized. This is because there is an exchange of stability, as indicated by the arrows. Thus, there is a tradeoff: w, and thus the mean firing rate of the memory neurons, can be increased by shifting the dashed line up or to the right , but eventually the background becomes destabilized. Shifting the dashed line to the left, on the other hand, will eventually eliminate the solution at w > 0, destroying all attractors but the background. For fixed load parameter Ct, fraction of neurons involved in a memory, f, and degree of connectivity, c, there are three parameters that have a large effect on the location of the equilibria in Fig. 1: the gain, {3, the clipped memory strength, fe, and the degree of heterogeneity in individual neurons, Bo. The effect of the first two can be seen in Fig. 2, which shows a stability plot in the f-{3 plane, determined by numerically solving the the equations Tdvi/dt = ?i - Vi (see Eq. (2)). The filled circles indicate regions where memories were embedded without destabilizing the background, open circles indicate regions where no memories could be embedded, and xs indicate regions where the background was unstable. As discussed above, fe becomes approximately independent of the strength of the memories, f, when f becomes large. This is seen in Fig. 2A, in which network behavior stabilizes when f becomes larger than about 4; increasing f beyond 8 would, presumably, produce no surprises. There is some sensitivity to gain, (3: when f > 4, memories co-existed with a stable background for (3 in a ?15% range. Although not shown, the same was true of eo: increasing it by about 20% eliminated the attractors; decreasing it by the same amount destabilized the background. However, more detailed analysis indicates that the stability region gets larger as the number of neurons in the network, N, increases. This is because fluctuations destabilize the background, and those fluctuations fall off as N - 1 / 2 . A E:o 70 11111",1 11 000000000000000 o 2 B '.2[\momo N !:S 35 ~ I o background ???? 4 0 0 ~ 4 8 E Figure 2: A. Stability diagram, found by solving the set of equations Tdv;/dt = cPi - Vi with the argument of cPi given in Eq. (2). Filled circles: memories co-exist with a stable background (also outlined with solid lines); open circles: memories could not be embedded; x s: background was unstable. The average background rate, when the background was stable, was around 3 Hz. The network parameters were eo = 6, Xo = 1.5, a = 0.5, c = 0.3, 0: = 2.5%, and 8w = 0.3. 2000 neurons were used in the simulations. These parameters led to an effective gain, pl /2 (3f c , of about 10, which is consistent with the gain in large networks in which each neuron receives "-'5-10,000 inputs. B . Plot of firing rate of memory neurons , m, when the memory was active (upper trace) and not active (lower trace) versus f at (3 = 2. 4 Discussion The main outcome of this analysis is that attractors can co-exist with a stable background when neurons have generic threshold-linear gain functions, so long as the sparse coding limit is avoided. The parameter regime for this co-existence is much larger than for attractor networks that operate in the sparse coding limit [2,23]. While these results are encouraging, they do not definitively establishing t hat attractors can exist in realistic networks. Future work must include inhibitory neurons , incorporate a much larger exploration of parameter space to ensure that the results are robust , and ultimately involve simulations with spiking neurons. 5 Acknowledgements This work was supported by NIMH grant #R01 MH62447. References [1] J.J. Hopfield. Neural networks and physical systems with emergent collective computational abilities. Proc. Natl. Acad. Sci ., 79:2554- 2558, 1982. [2] N. BruneI. Persistent activity and the single-cell frequency-current curve in a cortical network model. Network: Computation in Neural Systems, 11:261- 280, 2000. [3] P.E. Latham and S.N. Nirenberg. Intrinsic dynamics in cultured neuronal networks. Soc . Neuroscience Abstract, 25:2259, 1999. [4] J.M. Fuster and G.E. Alexander. Science, 173:652- 654, 1971. Neuron activity related to short-term memory. [5] Y. Miyashita. Inferior temporal cortex: where visual perception meets memory. Annu R ev Neurosci, 16:245- 263 , 1993. [6] P.S. Goldman-Rakic. Cellular basis of working memory. Neuron, 14:477- 485 , 1995. [7] R Romo, C.D. Brody, A. Hernandez , and L. Lemus. Neuronal correlates of parametric working memory in the prefrontal cortex. Nature , 399:470- 473, 1999. [8] C.D. Gilbert. Laminar differences in receptive field properties of cells in cat primary visual cortex. J. Physiol. , 268:391- 421 , 1977. [9] Y. Lamour, P. Dutar, and A. Jobert. Cerebral neorcortical neurons in the aged rat: spontaneous activity, properties of pyramidal tract neurons and effect of acetylcholine and cholinergic drugs. N euroscience, 16:835- 844, 1985. [10] M.B. Szente, A. Baranyi, and C.D. Woody. Intracellular injection of apamin reduces a slow potassium current mediating afterhyperpolarizations and IPSPs in neocortical neurons of cats. Brain Res. , 461:64- 74, 1988. [11] I. Salimi, H.H. Webster, and RW. Dykes. Neuronal activity in normal and deafferented forelimb somatosensory cortex of the awake cat . Brain Res., 656:263- 273, 1994. [12] J.F. Herrero and P.M. Headley. Cutaneous responsiveness of lumbar spinal neurons in awake and halothane-anesthetized sheep. J. N europhysiol. , 74:1549- 1562, 1997. [13] K. Ochi and J.J. Eggermont. Effects of quinine on neural activity in cat primary auditory cortex. Hear. Res., 105:105- 18, 1997. [14] P.E. Latham, B.J. Richmond, P.G. Nelson, and S.N. Nirenberg. Intrinsic dynamics in neuronal networks. I. Theory. J. Neurophysiol., 83:808- 827, 2000. [15] M.V. Tsodyks and M.V. Feigel'man. The enhanced storage capacity in neural networks with low activity level. Europhys. Lett. , 6:101- 105, 1988. [16] A. Treves. Mean-field analysis of neuronal spike dynamics. Network, 4:259- 284, 1993. [17] O. Shriki, D. Hansel , and H. Sompolonski. Modeling neuronal networks in cortex by rate models using the current-frequency response properties of cortical cells. Soc . Neurosci ence Abstract, 24:143 , 1998. [18] C. van Vreeswijk and H. Sompolinsky. Chaos in neuronal networks with balanced excitatory and inhibitory activity. Science, 274: 1724- 1726, 1996. [19] C. van Vreeswijk and H. Sompolinsky. Chaotic balanced state in a model of cortical circuits. Neural Comput., 10:1321- 1371 , 1998. [20] H. Sompolinsky. Neural networks with nonlinear synapses and a static noise. Phys. Rev. A, 34:2571- 2574, 1986. [21] J. Hertz , A. Krogh, and RG. Palmer. Introduction to the th eory of neural computation. Addison Wesley, Redwood City, CA, 1991. [22] A.N. Burkitt. Retrieval properties of attractor neural that obey Dale's law using a self-consistent signal-to-noise analysis. Network: Computation in Neural Systems, 7:517- 531 , 1996. [23] D.J. Amit and N. BruneI. Dynamics of a recurrent network of spiking neurons before and following learning. Network, 8:373- 404, 1997.
2056 |@word open:3 simulation:5 solid:7 reduction:1 moment:1 exclusively:1 neurobio:1 z2:1 current:3 must:5 written:3 physiol:1 realistic:7 webster:1 plot:2 nervous:2 plane:2 short:1 location:1 sigmoidal:1 along:1 constructed:1 become:1 persistent:1 behavior:4 brain:2 decreasing:1 goldman:1 encouraging:1 increasing:3 becomes:7 provided:1 totally:1 circuit:1 feigel:1 pel:2 what:2 interpreted:1 temporal:2 deafferented:1 unit:1 grant:1 positive:2 before:1 dropped:1 understood:1 limit:10 consequence:1 acad:1 establishing:1 meet:1 firing:27 solely:1 approximately:1 fluctuation:2 black:1 hernandez:1 co:6 dyke:1 palmer:1 range:1 spontaneously:1 chaotic:1 drug:1 destabilizing:2 significantly:1 pre:1 word:1 refers:1 get:1 cannot:1 superlinear:1 bend:1 storage:1 gilbert:1 destroying:1 romo:1 focused:1 stabilizing:1 deriving:1 embedding:1 population:4 stability:8 variation:1 cultured:1 spontaneous:1 enhanced:1 substrate:1 homogeneous:1 element:2 observed:2 role:1 capture:2 tsodyks:1 region:4 sompolinsky:3 decrease:1 rescaled:2 balanced:3 vanishes:1 complexity:1 nimh:1 dynamic:5 jobert:1 ultimately:1 solving:3 f2:1 completely:1 basis:1 neurophysiol:1 hopfield:3 emergent:1 cat:4 derivation:1 effective:2 outcome:1 europhys:1 whose:2 larger:4 nirenberg:2 ability:2 superscript:1 associative:4 vanishingly:1 strengthening:2 jij:5 inserting:1 relevant:1 description:1 getting:1 los:2 potassium:1 requirement:2 produce:2 tract:1 coupling:3 recurrent:2 ij:5 eq:27 krogh:1 soc:2 implemented:1 salimi:1 implies:1 indicate:6 somatosensory:1 direction:3 stabilization:2 exploration:1 exchange:2 quinine:1 pl:1 around:2 considered:1 normal:1 exp:2 presumably:1 equilibrium:11 shriki:1 stabilizes:1 a2:1 proc:1 iw:1 hansel:1 autapses:1 city:1 halothane:1 gaussian:5 always:1 super:1 acetylcholine:1 derived:1 consistently:1 indicates:1 lemus:1 richmond:1 contrast:1 typically:1 unlikely:2 eliminate:1 wij:2 among:3 exponent:1 development:1 field:10 f3:1 never:1 eliminated:1 represents:1 hardest:1 look:1 nearly:1 future:1 simplify:1 destabilized:3 few:3 ve:3 individual:1 attractor:20 conductance:2 brunei:2 highly:1 investigate:1 possibility:1 cholinergic:1 sheep:1 bracket:1 tj:3 natl:1 culture:1 filled:2 circle:4 plotted:1 re:3 theoretical:2 increased:1 modeling:1 ence:1 clipping:2 uniform:1 examining:1 too:2 accomplish:1 sensitivity:1 bu:1 off:1 connectivity:5 again:1 containing:1 prefrontal:1 derivative:1 return:1 li:1 potential:1 vei:2 coding:8 stabilize:1 vi:8 piece:1 break:2 doing:1 analyze:2 complicated:1 vivo:1 b6:1 square:3 variance:6 ofthe:1 ago:1 synapsis:1 fo:1 phys:1 synaptic:3 definition:2 frequency:2 involved:7 obvious:1 e2:1 associated:1 static:1 gain:16 auditory:1 recall:2 wesley:1 dt:3 follow:2 sustain:1 response:1 done:1 though:2 generality:1 implicit:1 working:3 hand:1 horizontal:2 receives:1 ei:5 nonlinear:1 indicated:1 perhaps:1 effect:5 normalized:1 y2:3 true:1 evolution:3 rakic:1 self:2 inferior:1 steady:1 excitation:1 rat:1 neocortical:1 latham:3 cp:4 meaning:2 chaos:1 spiking:2 physical:1 regulates:1 woody:1 spinal:1 cerebral:1 analog:1 discussed:2 numerically:1 ai:1 fk:4 outlined:1 similarly:1 nonlinearity:1 stable:15 cortex:6 attracting:1 inhibition:1 showed:1 prime:2 seen:3 responsiveness:1 greater:1 cii:1 eo:2 determine:2 dashed:7 signal:2 ii:1 multiple:4 full:2 reduces:2 af:2 believed:1 long:2 retrieval:1 prescription:1 post:1 equally:1 ipsps:1 involving:1 redefinition:1 essentially:1 represent:2 synaptically:1 cell:4 background:22 diagram:1 aged:1 pyramidal:1 zw:1 rest:1 eliminates:1 operate:1 probably:1 hz:4 virtually:1 flow:2 ee:3 near:1 intermediate:1 easy:1 enough:3 affect:1 competing:1 reduce:1 cn:1 tradeoff:1 angeles:2 shift:1 whether:2 expression:2 action:1 cpt:2 generally:1 useful:1 clear:1 detailed:1 involve:1 amount:1 band:1 rw:1 reduced:1 http:1 eory:1 exist:5 inhibitory:12 neuroscience:1 write:2 threshold:4 destabilize:1 falling:1 drawn:1 ce:1 fraction:4 sum:1 angle:1 powerful:1 europhysiol:1 clipped:3 almost:1 throughout:1 reasonable:2 brody:1 fl:1 pay:1 ct:1 simplification:2 existed:1 laminar:1 activity:9 strength:7 awake:2 x2:1 ucla:2 argument:2 injection:1 relatively:1 department:1 hertz:1 describes:1 rev:1 thresholdlinear:1 cutaneous:1 xo:3 taken:1 equation:17 pin:1 eventually:2 mechanism:1 vreeswijk:2 addison:1 letting:3 studying:1 apply:1 obey:1 away:1 generic:1 robustly:1 alternative:1 hat:1 existence:2 denotes:1 include:1 ensure:1 graphical:1 build:1 eggermont:1 amit:1 r01:1 bl:1 question:1 spike:1 parametric:1 receptive:1 dependence:2 primary:2 exhibit:2 separating:1 capacity:2 sci:1 participate:2 nelson:1 unstable:5 trivial:1 reason:1 cellular:1 assuming:1 providing:1 difficult:2 steep:1 cij:1 fe:2 potentially:2 mediating:1 trace:2 negative:3 collective:1 perform:1 allowing:1 upper:2 vertical:2 neuron:50 situation:1 neurobiology:1 communication:1 heterogeneity:1 redwood:1 lb:2 treves:1 inferred:1 required:1 connection:2 california:1 miyashita:1 beyond:1 below:3 pattern:2 ev:2 perception:1 regime:3 hear:1 saturation:1 max:1 memory:34 including:2 shifting:2 overlap:3 difficulty:2 treated:1 coupled:1 acknowledgement:1 relative:1 law:1 embedded:4 loss:1 sublinear:1 proportional:1 versus:1 degree:3 consistent:2 principle:1 uncorrelated:1 excitatory:9 surprisingly:1 supported:1 aij:1 understand:1 fall:1 definitively:1 sparse:8 anesthetized:1 distributed:1 van:2 feedback:5 curve:2 cortical:3 valid:1 lett:1 dale:1 simplified:1 avoided:1 ec:3 far:1 correlate:1 approximate:1 keep:1 active:5 xi:3 fuster:1 decade:1 euroscience:1 nature:1 zk:1 robust:2 ca:2 vj:1 main:5 linearly:1 arrow:5 neurosci:2 noise:6 intracellular:1 neuronal:13 fig:8 cpi:2 burkitt:1 definiteness:1 slow:1 sub:5 comput:1 lie:2 down:1 saturate:1 annu:1 load:2 x:1 intrinsic:2 surprise:1 vii:1 rg:1 intersection:3 led:1 simply:1 visual:2 saturating:1 bo:1 corresponds:4 determines:1 chance:1 goal:1 endeavor:1 marked:1 consequently:1 price:1 man:1 change:2 determined:2 except:2 operates:1 experimental:1 la:3 intact:1 indicating:1 support:1 arises:1 alexander:1 incorporate:1
1,159
2,057
Matching Free Trees with Replicator Equations Marcello Pelillo Dipartimento di Informatica Universit`a Ca? Foscari di Venezia Via Torino 155, 30172 Venezia Mestre, Italy E-mail: pelillo@dsi.unive.it Abstract Motivated by our recent work on rooted tree matching, in this paper we provide a solution to the problem of matching two free (i.e., unrooted) trees by constructing an association graph whose maximal cliques are in one-to-one correspondence with maximal common subtrees. We then solve the problem using simple replicator dynamics from evolutionary game theory. Experiments on hundreds of uniformly random trees are presented. The results are impressive: despite the inherent inability of these simple dynamics to escape from local optima, they always returned a globally optimal solution. 1 Introduction Graph matching is a classic problem in computer vision and pattern recognition, instances of which arise in areas as diverse as object recognition, motion and stereo analysis [1]. In many problems (e.g., [2, 11, 19]) the graphs at hand have a peculiar structure: they are connected and acyclic, i.e. they are free trees. Note that, unlike ?rooted? trees, in free trees there is no distinguished node playing the role of the root, and hence no hierarchy is imposed on them. Standard graph matching techniques, such as [8], yield solutions that are not constrained to preserve connectedness and hence cannot be applied to free trees. A classic approach to solving the graph matching problem consists of transforming it into the equivalent problem of finding a maximum clique in an auxiliary graph structure, known as the association graph [1]. This framework is attractive because it casts graph matching as a pure graph-theoretic problem, for which a solid theory and powerful algorithms have been developed. Note that, although the maximum clique problem is known to be hard, powerful heuristics exist which efficiently find good approximate solutions [4].  Motivated by our recent work on rooted tree matching [15], in this paper we propose a solution to the free tree matching problem by providing a straightforward way of deriving an association graph from two free trees. We prove that in the new formulation there is a one-to-one correspondence between maximal (maximum) cliques in the derived association graph and maximal (maximum) subtree isomorphisms. As an obvious corollary, the computational complexity of finding a maximum clique in such graphs is therefore the same as the subtree isomorphism problem, which is known to be polynomial in the number of nodes [7]. Following [13, 15], we use a recent generalization of the Motzkin-Straus theorem [12] to formulate the maximum clique problem as a quadratic programming problem. To (approximately) solve it we employ replicator equations, a class of simple continuous- and discretetime dynamical systems developed and studied in evolutionary game theory [10, 17]. We illustrate the power of the approach via experiments on hundreds of (uniformly) random trees. The results are impressive: despite the counter-intuitive maximum clique formulation of the tree matching problem, and the inherent inability of these simple dynamics to escape from local optima, they always found a globally optimal solution. 2 Subtree isomorphisms and maximal cliques         "!# $%   if   & !  "!  '  otherwise ( The degree of a node , denoted )+*-,./ 0 , is the number of nodes% adjacent to it. A path is any sequence of distinct nodes 012 435(2(-(6 87 such that for all 9: (2(2(; , =< 3>?  ; in this case, the length of the path is  . If 01@A B7 the path is called a cycle. A graph is said to be connected if any two nodes are joined by a path. The distance between two nodes and  , denoted by C8/ 5;+ , is the length of the shortest path joining them (by convention C8/ 5;+ DFE , if there is no such path). Given a subset of nodes GIHJ , the induced subgraph LK GNM is the graph having G as its node set, and two nodes are adjacent in LK GNM if and only if they are adjacent in . A connected graph with no cycles is called a free tree, Let be a graph, where is the set of nodes and is the set of (undirected) edges. The order of is the number of nodes in , while its size is the number of edges. Two nodes are said to be adjacent (denoted ) if they are connected by an edge. The adjacency matrix of is the symmetric matrix defined as or simply a tree. Trees have a number of interesting properties. One which turns out to be very useful for our characterization is that in a tree any two nodes are connected by a unique path. J.QS TQ# be two trees. Any bijection UWVYX 3[Z X\Q , with called a subtree isomorphism if it preserves both the adjacency Xrelationships [3OH]3 8J3 and 3 between XP Q 3 H?and Q the, O0isQRnodes and the connectedness of the matched subgraphs. Formally, this means that, given 5;[^X_3 , we have W if and only if U5/ 0 ]U= and, in addition, the induced subgraphs O 3 K X 3 M and O0Q`K XaQ2M are connected. A subtree isomorphism is maximal if there is no other subtree isomorphism U4bcV0X^3 b Z X^Q b with XR3 a strict subset of X3 b , and maximum if XR3 has largest cardinality. The maximal (maximum) subtree isomorphism problem is to find a maximal (maximum) subtree isomorphism between two Let trees. A word of caution about terminology is in order here. Despite name similarity, we are not addressing the so-called subtree isomorphism problem, which consists of determining whether a given tree is isomorphic to a subtree of a larger one. In fact, we are dealing with a generalization thereof, the maximum common subtree problem, which consists of determining the largest isomorphic subtrees of two given trees. We shall continue to use our own terminology, however, as it emphasizes the role of the isomorphism . U The free tree association graph (FTAG) of two trees Od3eB3fP N3- and O Q g  h Q  Q is the graph i j where W 3 @.Q (1) and, for any two nodes / 5;kN and /8 l in  , we have (2) =  kN /8 l nmoC.=   pWC8/kqPlr ( Note that this definition of the association graph is stronger than the standard one used for matching arbitrary relational structures [1]. A subset of vertices of is said to be a clique if all its nodes are mutually adjacent. A maximal clique is one which is not contained in any larger clique, while a maximum clique is a clique having largest cardinality. The maximum clique problem is to find a maximum clique of . The following theorem, which is the basis of the work reported here, establishes a one-toone correspondence between the maximum subtree isomorphism problem and the maximum clique problem. Theorem 1 Any maximal (maximum) subtree isomorphism between two trees induces a maximal (maximum) clique in the corresponding FTAG, and vice versa. be a maximal subtree isomorphism between trees On3 U&h:VYPX_ 3 denote Z X Q the Q O  corresponding FTAG. Let G  H  be defined as GU  maps r=the  U5path / 0 ; >between Vr X any the definition of a subtree isomorphism it follows that 3 . From two nodes   X 3 onto the path joining U5= B and U= . This clearly implies that C.=   \C.hU= B PU5/+ ; for all ? X_3 , and therefore G  is a clique. Trivially, G  is a maximal clique because U is maximal, and this proves the first Proof (outline). Let and , and let part of the theorem. maximal clique of , and let X 3  7 HW   isQ . aDefine for - 43#all9dY ; B% 7 (- (2H&(; .BGFrom 3 andr the X = Q 3 definition  ;k -3 k> 3f4 of#0/; thek 7 7 ;kFTAG UVrXR3 Z X Q as U5/  pWk  , and the hypothesis that G is a clique, it Suppose now that U X 3 U XaQ G is simple to see that is a one-to-one and onto correspondence between and , which trivially preserves the adjacency relationships between nodes. The fact that is a maximal isomorphism is a straightforward consequence of the maximality of . X 3 XaQ O 3 K X 3 M O0Q K XaQ2M X[3 OY3 On3n3(-(2K XL( 3 M g !    1 B % (2(-(   a       X  1  3n(2(2(  A k !   k   !  ! OQ C.=k  ;k ! > r    2C8 /   G NH  C8/k ;k  C.=   : C8/ 5 8 C8 ;+ To conclude the proof we have to show that the subgraphs that we obtain when we restrict ourselves to and , i.e. and , are trees, and this is equivalent to showing that they are connected. Suppose by contradiction that this is not the case, and let be two nodes which are not joined by a path in . Since both and are nodes of , however, there must exist a path joining them in . Let , for some , be a node on this path which is not in . Moreover, let be the -th node on the path which joins and in (remember that , and hence ). It is easy to show that the set is a clique, thereby contradicting the hypothesis that is a maximal clique. This can be proved by exploiting the obvious fact that if is a node on the path joining any two nodes and , then . ! O  3  !  3 k k! G The ?maximum? part of the statement is proved similarly. h:P   ! ^      C ! The FTAG is readily derived by using a classical representation for graphs, i.e., the socalled distance matrix which, for an arbitrary graph of order , is the matrix where , the distance between nodes and . Efficient, classical algorithms are available for obtaining such a matrix [6]. Note also that the distance matrix of a graph can easily be constructed from its adjacency matrix . In fact, denoting by the -th entry of the matrix , the -th power of , we have that equals the least for which (there must be such an since a tree is connected).   = C "! C "! ?C8/   !  7 !  /9 !  7 '  7  !#" 3 Matching free trees with replicator dynamics 7   j be an arbitrary graph of order  , and let $Y7 denote the standard simplex of $07L% &  IR7 (V ' b &^ % and * ) ' Y9d % (2(2(;  Let IR : ' G & where is the vector whose components equal 1, and a prime denotes transposition. Given a subset of vertices of , we will denote by  its characteristic vector which is the point in defined as $7 $ %  G   if p9 @G '  otherwise   where G denotes the cardinality of G . Now, consider the following quadratic function %    &Y  & b   &   & b & (3) where   g=  ! is the adjacency matrix of . The following theorem, recently proved by Bomze [3], expands on the Motzkin-Straus theorem [12], a remarkable result which es  tablishes a connection between the maximum clique problem and quadratic programming. G G $ 7 && Theorem 2 Let be a subset of vertices of a graph , and let  be its characteristic vector. Then, is a maximal (maximum) clique of if and only if is a local (global)   maximizer in . Moreover, all local (and hence global) maximizers of in are strict and are characteristic vectors of maximal cliques of .  $7  $07 Unlike the original Motzkin-Straus formulation, which is plagued by the presence  of ?spuon rious? solutions [14], the previous result guarantees us that all maximizers of are strict, and are characteristic vectors of maximal/maximum cliques in . In a formal sense, therefore,  a one-to-one correspondence exists between maximal cliques and local maximizers of in on the one hand, and maximum cliques and global maximizers on the other hand.  $7 A/kN"!   7 .P ;  . P ;   ;  !  3 +!  ; !r ;   @ We now turn our attention to a class of simple dynamical systems that we use for solving our quadratic optimization problem. Let be a non-negative real-valued matrix, and consider the following continuous-time dynamical system: (4) where a dot signifies derivative with respect to time, and its discrete-time counterpart: where  % 8    ; P    ; 8   :  !7  3 !  ; !  ;  7   ; : !  3 k  ! !  ; ( (5) (6) Both (4) and (5) are called replicator equations in evolutionary game theory, since they are used to model evolution over time of relative frequencies of interacting, self-replicating entities [10, 17]. It is readily seen that the simplex is invariant under these dynamics, which means that every trajectory starting in will remain in for all future times, and their stationary points coincide. $7 $7 $47 We are now interested in the dynamical properties of replicator dynamics; it is these properties that will allow us to solve our original tree matching problem. The following result is known in mathematical biology as the fundamental theorem of natural selection [10, 17] and, in its original form, traces back to R. A. Fisher.  ib &b & Theorem 3 If is strictly increasing along any non then the function constant trajectory under both continuous-time (4) and discrete-time (5) replicator dynamics. Furthermore, any such trajectory converges to a stationary point. Finally, a vector is asymptotically stable under (4) and (5) if and only if is a strict local maximizer of on . & &4b $ 7 & $07 & In light of their dynamical properties, replicator equations naturally suggest themselves as a simple heuristic for solving the maximal subtree isomorphism problem. Indeed, let and be two free trees, and let denote the adjacency matrix of their FTAG . By letting O 3 Ih 3  3 O0Q I.Q`P TQ- W     %   (7)  where is the identity matrix, we know that the replicator dynamical systems (4)  and (5), starting from an arbitrary initial state, will iteratively maximize the function defined in (3) over the simplex and will eventually converge with probability 1 to a strict local maximizer which, by virtue of Theorem 2, will then correspond to the characteristic vector of a maximal clique in the association graph. As stated in Theorem 1, this will in turn induce a maximal subtree isomorphism between and . Clearly, in theory there is no guarantee that the converged solution will be a global maximizer of , and therefore that it will induce a maximum isomorphism between the two original trees, but see below. O3 O0Q  Recently, there has been much interest around the following exponential version of replicator equations, which arises as a model of evolution guided by imitation [9, 10, 17]:   ;    ;   %    !7  3 !  ;       (8) where  is a positive constant. As  tends to 0, the orbits of this dynamics approach those of the standard, ?first-order? replicator model (4), slowed down by the factor  ; moreover, for large values of  the model approximates the so-called ?best-reply? dynamics [9, 10]. A customary way of discretizing equation (8) is given by the following difference equations:    % :  !7   3  ; !  ;        ( (9) From a computational perspective, exponential replicator dynamics are particularly attractive as they may be considerably faster and even more accurate than the standard, first-order model (see [13] and the experiments reported in the next section). 4 Results and conclusions We tested our algorithms over large random trees. Random structures represent a useful benchmark not only because they are not constrained to any particular application, but also because it is simple to replicate experiments and hence to make comparisons with other algorithms. In this series of experiments, the following protocol was used. A hundred 100-node free trees were generated uniformly at random using a procedure described by Wilf in [18]. Then, each such tree was subject to a corruption process which consisted of randomly deleting a fraction of its nodes (in fact, the to-be-deleted nodes were constrained to be the terminal ones, otherwise the resulting graph would have been disconnected), thereby obtaining a tree isomorphic to a proper subtree of the original one. Various levels of corruption (i.e., percentage of node deletion) were used, namely 2%, 10%, 20%, 30% and 40%. This means that the order of the pruned trees ranged from 98 to 60. Overall, therefore, 500 pairs of trees were obtained, for each of which the corresponding FTAG was constructed as described in Section 2. To keep the order of the association graph as low as possible, its vertex set was constructed as follows:  %r=  kN @ b @ b V)+*2,8/ 0  ) *2,B/kN    b , the edge set being defined as in (2). It is straightforward to see assuming  b that when the first tree is isomorphic to a subtree of the second, Theorem 1 continues to    hold. This simple heuristic may significantly reduce the dimensionality of the search space. We also performed some experiments with unpruned FTAG?s but no significant difference in performance was noticed apart, of course, heavier memory requirements.  % ' Both the discrete-time first-order dynamics (5) and its exponential counterpart (9) (with ) were used. The algorithms were started from the simplex barycenter and stopped   ) was found or the distance when either a maximal clique (i.e., a local maximizer of between two successive points was smaller than a fixed threshold. In the latter case the converged vector was randomly perturbed, and the algorithms restarted from the perturbed point. Note that this situation corresponds to convergence to a saddle point.  After convergence, we calculated the proportion of matched nodes, i.e., the ratio between the cardinality of the clique found and the order of the smaller subtree, and then we averaged. Figure 1(a) shows the results obtained using the linear dynamics (5) as a function of the corruption level. As can be seen, the algorithm was always able to find a correct maximum isomorphism, i.e. a maximum clique in the FTAG. Figure 1(b) plots the corresponding (average) CPU time taken by the processes, with corresponding error bars (simulations were performed on a machine equipped with a 350MHz AMDK6-2 processor). In Figure 2, the results pertaining to the exponential dynamics (8) are shown. In terms of solution?s quality the algorithm performed exactly as its linear counterpart, but this time it was dramatically faster. This confirms earlier results reported in [13]. Before concluding, we note that our approach can easily be extended to tackle the problem of matching attributed (free) trees. In this case, the attributes result in weights being placed on the nodes of the association graph, and a conversion of the maximum clique problem to a maximum weight clique problem [15, 5]. Moreover, it is straightforward to formulate errortolerant versions of our framework along the lines suggested in [16] for rooted attributed trees, where many-to-many node correspondences are allowed. All this will be the subject of future investigations. Finally, we think that the results presented in this paper (together with those reported in [13, 15]) raise intriguing questions concerning the connections between (standard) notions of computational complexity and the ?elusiveness? of global optima in continuous settings. Acknowledgments. The author would like to thank M. Zuin for his support in performing the experiments. References [1] D. H. Ballard and C. M. Brown. Computer Vision. Prentice-Hall, Englewood Cliffs, NJ, 1982. [2] H. Blum and R. N. Nagel. Shape description using weighted symmetric axis features. Pattern Recognition, 10:167?180, 1978. [3] I. M. Bomze. Evolution towards the maximum clique. J. Glob. Optim., 10:143?164, 1997. [4] I. M. Bomze, M. Budinich, P. M. Pardalos, and M. Pelillo. The maximum clique problem. In D.-Z. Du and P. M. Pardalos, editors, Handbook of Combinatorial Optimization (Suppl. Vol. A), pages 1?74. Kluwer, Boston, MA, 1999. [5] I. M. Bomze, M. Pelillo, and V. Stix. Approximating the maximum weight clique using replicator dynamics. IEEE Trans. Neural Networks, 11(6):1228?1241, 2000. Figure 1: Results obtained over 100-node random trees with various levels of corruption, using the first-order dynamics (5). Top: Percentage of correct matches. Bottom: Average computational time taken by the replicator equations. [6] T. H. Cormen, C. E. Leiserson, and R. L. Rivest. Introduction to Algorithms. MIT Press, Cambridge, MA, 1990. [7] M. R. Garey and D. S. Johnson. Computers and Intractability: A Guide to the Theory of NPCompleteness. W. H. Freeman, San Francisco, CA, 1979. [8] S. Gold and A. Rangarajan. A graduated assignment algorithm for graph matching. IEEE Trans. Pattern Anal. Machine Intell. 18:377-388, 1996. [9] J. Hofbauer. Imitation dynamics for games. Collegium Budapest, preprint, 1995. [10] J. Hofbauer and K. Sigmund. Evolutionary Games and Population Dynamics. Cambridge University Press, Cambridge, UK, 1998. [11] T.-L. Liu, D. Geiger, and R. V. Kohn. Representation and self-similarity of shapes. In Proc. ICCV?98?6th Int. Conf. Computer Vision, pages 1129?1135, Bombay, India, 1998. 500 1000 1500 95 2500 100 2000 Percentage Average CPU of correct time (inmatches secs) 3000 [12] T. S. Motzkin and E. G. Straus. Maxima for graphs and a new proof of a theorem of Tur?an. Canad. J. Math., 17:533?540, 1965.   Figure 2: Results obtained over 100-node random trees with various levels of corruption, using the exponential dynamics (9) with . Top: Percentage of correct matches. Bottom: Average computational time taken by the replicator equations. 100 200 95 100 300 Percentage Average CPU of correct time (inmatches secs) 400 [13] M. Pelillo. Replicator equations, maximal cliques, and graph isomorphism. Neural Computation, 11(8):2023?2045, 1999. [14] M. Pelillo and A. Jagota. Feasible and infeasible maxima in a quadratic program for maximum clique. J. Artif. Neural Networks, 2:411?420, 1995. [15] M. Pelillo, K. Siddiqi, and S. W. Zucker. Matching hierarchical structures using association graphs. IEEE Trans. Pattern Anal. Machince Intell., 21(11):1105?1120, 1999. [16] M. Pelillo, K. Siddiqi, and S. W. Zucker. Many-to-many matching of attributed trees using association graphs and game dynamics. In C. Arcelli, L. P. Cordella, and G. Sanniti di Baja, editors, Visual Form 2001, pages 583?593. Springer, Berlin, 2001. [17] J. W. Weibull. Evolutionary Game Theory. MIT Press, Cambridge, MA, 1995. [18] H. Wilf. The uniform selection of free trees. J. Algorithms, 2:204?207, 1981. [19] S. C. Zhu and A. L. Yuille. FORMS: A flexible object recognition and modeling system. Int. J. Computer Vision, 20(3):187?212, 1996.
2057 |@word version:2 polynomial:1 proportion:1 stronger:1 replicate:1 hu:1 confirms:1 simulation:1 thereby:2 gnm:2 solid:1 initial:1 liu:1 series:1 denoting:1 optim:1 intriguing:1 must:2 readily:2 shape:2 plot:1 stationary:2 transposition:1 characterization:1 math:1 node:34 bijection:1 successive:1 mathematical:1 along:2 constructed:3 consists:3 prove:1 indeed:1 themselves:1 terminal:1 freeman:1 globally:2 p9:1 cpu:3 equipped:1 cardinality:4 increasing:1 matched:2 moreover:4 rivest:1 weibull:1 developed:2 caution:1 finding:2 nj:1 machince:1 guarantee:2 remember:1 bomze:4 every:1 expands:1 tackle:1 exactly:1 universit:1 uk:1 glob:1 unive:1 positive:1 before:1 local:8 tends:1 consequence:1 despite:3 joining:4 cliff:1 path:13 approximately:1 connectedness:2 studied:1 averaged:1 unique:1 acknowledgment:1 procedure:1 area:1 significantly:1 matching:17 word:1 induce:2 suggest:1 cannot:1 onto:2 selection:2 prentice:1 equivalent:2 imposed:1 map:1 straightforward:4 attention:1 starting:2 formulate:2 pure:1 subgraphs:3 gihj:1 q:1 contradiction:1 deriving:1 oh:1 his:1 classic:2 population:1 notion:1 hierarchy:1 suppose:2 programming:2 hypothesis:2 recognition:4 particularly:1 continues:1 bottom:2 role:2 preprint:1 connected:8 cycle:2 counter:1 tur:1 transforming:1 complexity:2 dynamic:19 raise:1 solving:3 yuille:1 basis:1 gu:1 easily:2 various:3 distinct:1 pertaining:1 venezia:2 whose:2 heuristic:3 larger:2 solve:3 valued:1 otherwise:3 think:1 sequence:1 propose:1 maximal:26 budapest:1 subgraph:1 gold:1 intuitive:1 description:1 exploiting:1 convergence:2 optimum:3 requirement:1 rangarajan:1 converges:1 object:2 illustrate:1 pelillo:8 auxiliary:1 implies:1 convention:1 guided:1 correct:5 attribute:1 pardalos:2 adjacency:6 generalization:2 investigation:1 dipartimento:1 strictly:1 hold:1 around:1 hall:1 plagued:1 proc:1 combinatorial:1 largest:3 vice:1 establishes:1 weighted:1 mit:2 clearly:2 always:3 thek:1 corollary:1 derived:2 sense:1 interested:1 overall:1 flexible:1 denoted:3 socalled:1 constrained:3 equal:2 having:2 biology:1 marcello:1 future:2 simplex:4 inherent:2 escape:2 employ:1 randomly:2 preserve:3 intell:2 ourselves:1 tq:2 interest:1 englewood:1 leiserson:1 light:1 subtrees:2 accurate:1 peculiar:1 edge:4 tree:44 orbit:1 toone:1 stopped:1 instance:1 earlier:1 modeling:1 bombay:1 mhz:1 assignment:1 signifies:1 vertex:4 addressing:1 subset:5 entry:1 hundred:3 kq:1 uniform:1 johnson:1 reported:4 kn:5 perturbed:2 considerably:1 fundamental:1 mestre:1 together:1 conf:1 derivative:1 sec:2 wk:1 int:2 bg:1 performed:3 root:1 ir:1 characteristic:5 efficiently:1 yield:1 correspond:1 emphasizes:1 trajectory:3 corruption:5 processor:1 converged:2 definition:3 frequency:1 obvious:2 straus:4 thereof:1 proof:3 di:3 naturally:1 attributed:3 garey:1 proved:3 dimensionality:1 back:1 y7:1 formulation:3 furthermore:1 reply:1 hand:3 maximizer:5 quality:1 artif:1 name:1 consisted:1 ranged:1 brown:1 counterpart:3 evolution:3 hence:5 symmetric:2 iteratively:1 attractive:2 adjacent:5 game:7 self:2 rooted:4 o3:1 outline:1 theoretic:1 motion:1 sigmund:1 recently:2 common:2 replicator:16 nh:1 association:11 approximates:1 kluwer:1 significant:1 versa:1 cambridge:4 trivially:2 similarly:1 replicating:1 dot:1 stable:1 zucker:2 impressive:2 similarity:2 own:1 recent:3 perspective:1 italy:1 apart:1 prime:1 discretizing:1 continue:1 jagota:1 seen:2 xaq:2 unrooted:1 converge:1 shortest:1 maximize:1 faster:2 match:2 concerning:1 j3:1 vision:4 represent:1 suppl:1 addition:1 unlike:2 strict:5 induced:2 subject:2 nagel:1 undirected:1 oq:1 presence:1 easy:1 b7:1 graduated:1 restrict:1 plr:1 reduce:1 whether:1 motivated:2 heavier:1 kohn:1 isomorphism:20 stereo:1 returned:1 dramatically:1 useful:2 u5:3 induces:1 siddiqi:2 informatica:1 discretetime:1 exist:2 percentage:5 vy:1 diverse:1 discrete:3 shall:1 vol:1 terminology:2 threshold:1 blum:1 deleted:1 graph:33 asymptotically:1 fraction:1 powerful:2 geiger:1 collegium:1 correspondence:6 quadratic:5 foscari:1 n3:1 c8:7 pruned:1 concluding:1 performing:1 disconnected:1 cormen:1 remain:1 smaller:2 slowed:1 invariant:1 iccv:1 taken:3 equation:10 mutually:1 turn:3 eventually:1 know:1 letting:1 available:1 hierarchical:1 distinguished:1 customary:1 original:5 denotes:2 top:2 hofbauer:2 prof:1 approximating:1 classical:2 torino:1 noticed:1 question:1 barycenter:1 canad:1 said:3 evolutionary:5 distance:5 thank:1 berlin:1 entity:1 mail:1 uwvyx:1 assuming:1 length:2 relationship:1 providing:1 ratio:1 fe:1 statement:1 trace:1 negative:1 stated:1 anal:2 proper:1 conversion:1 on3:1 benchmark:1 situation:1 relational:1 extended:1 interacting:1 arbitrary:4 cast:1 namely:1 pair:1 connection:2 deletion:1 trans:3 able:1 bar:1 suggested:1 dynamical:6 pattern:4 below:1 program:1 memory:1 deleting:1 power:2 natural:1 zhu:1 maximality:1 lk:2 started:1 axis:1 determining:2 relative:1 dsi:1 interesting:1 acyclic:1 remarkable:1 degree:1 unpruned:1 editor:2 intractability:1 playing:1 course:1 placed:1 free:14 infeasible:1 formal:1 allow:1 guide:1 india:1 calculated:1 author:1 coincide:1 san:1 approximate:1 keep:1 clique:40 dealing:1 global:5 handbook:1 conclude:1 francisco:1 imitation:2 continuous:4 search:1 ballard:1 ca:2 obtaining:2 du:1 constructing:1 protocol:1 arise:1 contradicting:1 allowed:1 join:1 vr:1 exponential:5 xl:1 ib:1 hw:1 theorem:13 down:1 showing:1 npcompleteness:1 virtue:1 maximizers:4 exists:1 subtree:20 boston:1 simply:1 saddle:1 visual:1 motzkin:4 contained:1 joined:2 restarted:1 springer:1 corresponds:1 ma:3 identity:1 towards:1 fisher:1 feasible:1 hard:1 wilf:2 uniformly:3 called:6 isomorphic:4 e:1 formally:1 support:1 latter:1 inability:2 arises:1 tested:1
1,160
2,058
K-Local Hyperplane and Convex Distance Nearest Neighbor Algorithms Pascal Vincent and Yoshua Bengio Dept. IRO, Universit?e de Montr?eal C.P. 6128, Montreal, Qc, H3C 3J7, Canada vincentp,bengioy @iro.umontreal.ca http://www.iro.umontreal.ca/ vincentp   Abstract Guided by an initial idea of building a complex (non linear) decision surface with maximal local margin in input space, we give a possible geometrical intuition as to why K-Nearest Neighbor (KNN) algorithms often perform more poorly than SVMs on classification tasks. We then propose modified K-Nearest Neighbor algorithms to overcome the perceived problem. The approach is similar in spirit to Tangent Distance, but with invariances inferred from the local neighborhood rather than prior knowledge. Experimental results on real world classification tasks suggest that the modified KNN algorithms often give a dramatic improvement over standard KNN and perform as well or better than SVMs. 1 Motivation The notion of margin for classification tasks has been largely popularized by the success of the Support Vector Machine (SVM) [2, 15] approach. The margin of SVMs has a nice geometric interpretation1: it can be defined informally as (twice) the smallest Euclidean distance between the decision surface and the closest training point. The decision surface produced by the original SVM algorithm is the hyperplane that maximizes this distance while still correctly separating the two classes. While the notion of keeping the largest possible safety margin between the decision surface and the data points seems very reasonable and intuitively appealing, questions arise when extending the approach to building more complex, non-linear decision surfaces. Non-linear SVMs usually use the ?kernel trick? to achieve their non-linearity. This conceptually corresponds to first mapping the input into a higher-dimensional feature space with some non-linear transformation and building a maximum-margin hyperplane (a linear decision surface) there. The ?trick? is that this mapping is never computed directly, but implicitly induced by a kernel. In this setting, the margin being maximized is still the smallest Euclidean distance between the decision surface and the training points, but this time measured in some strange, sometimes infinite dimensional, kernel-induced feature space rather than the original input space. It is less clear whether maximizing the margin in this new space, is meaningful in general (see [16]). 1 for the purpose of this discussion, we consider the original hard-margin SVM algorithm for two linearly separable classes. A different approach is to try and build a non-linear decision surface with maximal distance to the closest data point as measured directly in input space (as proposed in [14]). We could for instance restrict ourselves to a certain class of decision functions and try to find the function with maximal margin among this class. But let us take this even further. Extending the idea of building a correctly separating non-linear decision surface as far away as possible from the data points, we define the notion of local margin as the Euclidean distance, in input space, between a given point on the decision surface and the closest training point. Now would it be possible to find an algorithm that could produce a decision surface which correctly separates the classes and such that the local margin is everywhere maximal along its surface? Surprisingly, the plain old Nearest Neighbor algorithm (1NN) [5] does precisely this! So why does 1NN in practice often perform worse than SVMs? One typical explanation, is that it has too much capacity, compared to SVM, that the class of function it can produce is too rich. But, considering it has infinite capacity (VC-dimension), 1NN is still performing quite well... This study is an attempt to better understand what is happening, based on geometrical intuition, and to derive an improved Nearest Neighbor algorithm from this understanding. 2 Fixing a broken Nearest Neighbor algorithm 2.1 Setting and definitions  (the input space).      We are given corresponding     set  atraining    of  points    !  #"$ %"  '&(   and  )!their * )!* class label where is the + ,  number of different classes. The pairs are assumed to be samples drawn from an  ./ 0 unknown . Barring 76 class labels associated to 1 distribution 1 inputs, 4 5the *  duplicate  define a partition of  : let  32 each . " 97: The problem is to find that will generalize well on new .=a >decision 0  9 function 8  <; points drawn from . 8 should ideally minimize the expected classification.= error, 0  M O$P where ?A@ denotes the expectation with respect to i.e. minimize ?A@CB DFG'E HJILKNF T   S     & 9 M R denotes the indicator function, whose value is if 8 and DFG(E HJQKN and U otherwise. The setting is that of a classical classification problem in     In the previous and following discussion, we often refer to the concept of decision surface, 9 also known as decision The function 8 corresponding to a given de6 #boundary. W " *  algorithm  2 fines for any class two regions of the input space: the region V  6 * 9  5X6 and its complement  Y V . The decision surface for class is the ?boundary? 8 & * between those two regions, i.e. the contour of V , and can be seen as a Z Y dimensional manifold (a ?surface? in   ) possibly made of several disconnected components. For simplicity, when we mention the decision surface in our discussion we consider only the case of two class discrimination, in which there is a single decision surface.  1   that When we mention a test point, we mean a point  not belong to the training 9  does set  and for which the algorithm is to decide on a class 8 . By distance, we mean the usual Euclidean in input-space  distance  tween two points [ and \ will be written ] [ \ or alternatively ^[ Y  The distance between ba<cJdfe point g'h + and + a`single i  a set of points point of the set: ] _ ] . The  K-neighborhood jLk to is smallest.   of a test point *   The  K-c-neighborhood jmk to is smallest.  of a test point _ is the set of the l  is the set of l  . \(^ . The distance be- is the distance to the closest points of  whose distance points of  * whose distance By Nearest Neighbor algorithm (1NN) we mean the following algorithm: the class of a test  point is decided to be the same as the class of its closest neighbor in _ . By K-Nearest  Neighbor algorithm (KNN) we mean the following algorithm: the class of a test point is decided to be the same as the class appearing most frequently among the  K-neighborhood of . 2.2 The intuition Figure 1: A local view of the decision surface produced by the Nearest Neighbor (left) and SVM (center) algorithms, and how the Nearest Neighbor solution gets closer to the SVM solution in the limit, if the support for the density of each class is a manifold which can be considered locally linear (right). Figure 1 illustrates a possible intuition about why SVMs outperforms 1NNs when we have a finite number of samples. For classification tasks where the classes are considered to be mostly separable,2 we often like to think of each class as residing close to a lowerdimensional manifold (in the high dimensional input space) which can reasonably be considered locally linear3 . In the case of a finite number of samples, ?missing? samples would appear as ?holes? introducing artifacts in the decision surface produced by classical Nearest Neighbor algorithms. Thus the decision surface, while having the largest possible local margin with regard to the training points, is likely to have a poor small local margin with respect to yet unseen samples falling close to the locally linear manifold, and will thus result in poor generalization performance. This problem fundamentally remains with the K Nearest Neighbor (KNN) variant of the algorithm, but, as can be seen on the figure, it does not seem to affect the decision surface produced by SVMs (as the surface is constrained to a particular smooth form, a straight line or hyperplane in the case of linear SVMs). It is interesting to notice, if the assumption of locally linear class manifolds holds, how the 1NN solution approaches the SVM solution in the limit as we increase the number of samples. To fix this problem, the idea is to somehow fantasize the missing points, based on a local linear approximation of the manifold of each class. This leads to modified Nearest Neighbor algorithms described in the next sections.4 2 By ?mostly separable? we mean that the Bayes error is almost zero, and the optimal decision surface has not too many disconnected components. 3 i.e. each class has a probability density with a ?support? that is a lower-dimensional manifold, and with the probability quickly fading, away from this support. 4 Note that although we never generate the ?fantasy? points explicitly, the proposed algorithms are really equivalent to classical 1NN with fantasized points. 2.3 The basic algorithm  Given a test point , we are really interested in finding the closest neighbor, not among the training set  , but among an abstract, virtually enriched training set that would contain all the fantasized ?missing? points of the manifold of each6 class, locally approximated by an affine subspace. We shall thus consider, for each class  , the local affine subspace that passes& through the l points of the K-c neighborhood of . This affine subspace is typically l Y dimensional or less, and we will somewhat abusively call it the ?local hyperplane?.5 Formally, the local hyperplane can be defined as     * k  5 i i   k  )         N N  )   >)   *   . j k where  (1)   & , ) is to Another way to define this hyperplane, that gets rid of the constraint     take point within the hyperplane as an origin, for instance the centroid 6   a reference )   . This same hyperplane can then be expressed as  k N k  * k  5 ;Y    )  )  Y where  i i   )  k  ;Y         N  . (2) 6  Our modified nearest then associates a<c d point    whose 9  Aa$test  *   neighbor algorithm * g ] +to the * kclass k is closest to . Formally , where hyperplane 8 F  *  , ] k is logically called K-local Hyperplane Distance, hence the name K-local Hyperplane Distance Nearest Neighbor algorithm (HKNN in short). Computing, for each class ] 6 F  *  , k  e g !a!#c "d H QK ^  Y i ^ $ a<cJd  )   k  ;Y   + (3) Y % 'g & (*),+++ Y ++  N   ++ ++ amounts to solving a linear system in  , that can be easily expressed in matrix form as:  .- /0  /  1- /  )   (4) Y    )      where and  are Z dimensional column vectors,   ;Y  matrix whose columns are the  vectors defined earlier.7  k - , and  is a Z2 l Strictly speaking a hyperplane in an 3 dimensional input space is an 35476 affine subspace, while our ?local hyperplanes? can have fewer dimensions. 6 We could be using one of the 8 neighbors as the reference point, but this formulation with the centroid will prove useful later. 94 :7; 7 Actually there is an infinite number of solutions to this system since the are linearly dependent: remember that the initial formulation had an equality constraint and thus only 8<4=6 effective degrees of freedom. But we are interested 94 :7; in >@?BADCFEHGJK I ?BA LFL not in M so any solution will do. Alternatively, we can remove one of the from the system so that it has a unique solution. 5 2.4 Links with other paradigms The proposed  HKNN *   algorithm is very similar in spirit to the Tangent Distance Algok rithm [13]. can be seen as a tangent hyperplane representing a set of local di;Y  rections of transformation (any linear combination of the  vectors) that do not affect the class identity. These are invariances. The main difference is that in HKNN these invariances are inferred directly from the local neighborhood in the training set, whereas in Tangent Distance, they are based on prior knowledge. It should be interesting (and relatively easy) to combine both approaches for improved performance when prior knowledge is available. Previous work on nearest-neighbor variations based on other locally-defined metrics can be found in [12, 9, 6, 7], and is very much related to the more general paradigm of Local Learning Algorithms [3, 1, 10]. We should also mention close similarities between our approach and the recently proposed Local Linear Embedding [11] method for dimensionality reduction. The idea of fantasizing points around the training points in order to define the decision surface is also very close to methods based on estimating the class-conditional input density [14, 4]. Besides, it is interesting to look at HKNN from a different, less geometrical angle: it can be understood as choosing the class that achieves the best reconstruction (the smallest reconstruction error) of the test pattern through a linear combination of l particular prototypes of that class (the l neighbors). From this point of view, the algorithm is very similar to the Nearest Feature Line (NFL) [8] method. They differ in the fact that NFL considers all pairs of points for its search rather than the local l neighbors, thus looking at many (  ) & lines (i.e. 2 dimensional affine subspaces), rather than at a single l Y dimensional one. 3 Fixing the basic HKNN algorithm 3.1 Problem arising for large K One problem with the basic HKNN algorithm, as previously described, arises as we increase the value of l , i.e. the number of points considered in the neighborhood of the test point. In a typical high dimensional setting, exact colinearities between input patterns are rare, which means that as soon as l Z , any pattern of   (including nonsensical ones) can be produced by a linear combination of the l neighbors. The ?actual? dimensionality of the manifold may be much less than l . This is due to ?near-colinearities? di - /0 that producing rections associated to small eigenvalues of the covariance matrix are but noise, that can lead the algorithm to mistake those noise directions for ?invariances?, and may hurt its performance even for smaller values of l . Another related issue is that the linear approximation of the class manifold by a hyperplane is valid only locally, so we might want to restrict the ?fantasizing? of class members to a smaller region of the hyperplane. We considered two ways of dealing with these problems.8 3.2 The convex hull solution One way to avoid the above mentioned problems is to restrict ourselves to considering the convex hull of the neighbors, rather than the whole hyperplane they support (of which   the convex hull is a subset). This corresponds to adding a constraint of   U to equation (1). Unlike the problem of computing the distance to the hyperplane, the distance to the convex hull cannot be found by solving a simple linear system, but typically requires solving a quadratic programming problem (very similar to the one of SVMs). While this 8 A third interesting avenue, which : we did not have time to explore, would be to keep only the most relevant principal components of , ignoring those corresponding to small eigenvalues. is more complex to implement, it should be mentioned that the problems to be solved are of a relatively small dimension of order l , and that the time of the whole algorithm will very likely still be dominated by the search of the l nearest neighbors within each class. This algorithm will be referred to as K-local Convex Distance Nearest Neighbor Algorithm (CKNN in short). 3.3 The ?weight decay? penalty solution This consists in incorporating a penalty term to equation (3) to penalize large values of (i.e. it penalizes moving away from the centroid, especially in non essential directions):  - +  * k    a<g'cJ& (0d )+  Y )  Y  k  ;Y   +   k  (5) % ++ ++  N   N  ++   - /  ++   /   - /  )   The solution for  is given by solving the linear system D  Y 2 Z identity matrix. This is equation (4) with an additional diagonal term. where D  is the Z7T ] The resulting algorithm is a generalization of HKNN (basic HKNN corresponds to U ). 4 Experimental results  We performed a number of experiments, to highlight different properties of the algorithms: A first 2D toy example (see Figure 2) graphically illustrates the qualitative differences in the decision surfaces produced by KNN, linear SVM and CKNN. Table 1 gives quantitative results on two real-world digit OCR tasks, allowing to compare the performance of the different old and new algorithms. Figure 3 illustrates the problem arising with large l , mentioned in Section 3, and shows that the two proposed solutions: CKNN and HKNN with an added weight decay , allow to overcome it. In our final experiment, we wanted to see if the good performance of the new algorithms absolutely depended on having all the training points at hand, as this has a direct impact on speed. So we checked what performance we could get out of HKNN and CKNN when using only a small but representative subset of the training points, namely the set of support vectors found by a Gaussian Kernel SVM. The results obtained for MNIST are given in Table 2, and look very encouraging. HKNN appears to be able to perform as well or better than SVMs without requiring more data points than SVMs.    Figure 2: 2D illustration of the decision surfaces produced by KNN (left, K=1), linear SVM (middle), and CKNN (right, K=2). The ?holes? are again visible in KNN. CKNN doesn?t suffer from this, but keeps the objective of maximizing the margin locally. 5 Conclusion From a few geometrical intuitions, we have derived two modified versions of the KNN algorithm that look very promising. HKNN is especially attractive: it is very simple to implement on top of a KNN system, as it only requires the additional step of solving a small and simple linear system, and appears to greatly boost the performance of standard KNN even above the level of SVMs. The proposed algorithms share the advantages of KNN (no training required, ideal for fast adaptation, natural handling of the multi-class case) and its drawbacks (requires large memory, slow testing). However our latest result also indicate the possibility of substantially reducing the reference set in memory without loosing on accuracy. This suggests that the algorithm indeed captures essential information in the data, and that our initial intuition on the nature of the flaw of KNN may well be at least partially correct. References [1] C. G. Atkeson, A. W. Moore, and S. Schaal. Locally weighted learning. Artificial Intelligence Review, 1996. [2] B. Boser, I. Guyon, and V. Vapnik. An algorithm for optimal margin classifiers. In Fifth Annual Workshop on Computational Learning Theory, pages 144?152, Pittsburgh, 1992. [3] L. Bottou and V. Vapnik. Local learning algorithms. Neural Computation, 4(6):888?900, 1992. [4] Olivier Chapelle, Jason Weston, L?eon Bottou, and Vladimir Vapnik. Vicinal risk minimization. In T.K. Leen, T.G. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing Systems, volume 13, pages 416?422, 2001. [5] T.M. Cover and P.E. Hart. Nearest neighbor pattern classification. IEEE Transactions on Information Theory, 13(1):21?27, 1967. [6] J. Friedman. Flexible metric nearest neighbor classification. Technical Report 113, Stanford University Statistics Department, 1994. [7] Trevor Hastie and Robert Tibshirani. Discriminant adaptive nearest neighbor classification and regression. In David S. Touretzky, Michael C. Mozer, and Michael E. Hasselmo, editors, Advances in Neural Information Processing Systems, volume 8, pages 409?415. The MIT Press, 1996. [8] S.Z. Li and J.W. Lu. Face recognition using the nearest feature line method. IEEE Transactions on Neural Networks, 10(2):439?443, 1999. [9] J. Myles and D. Hand. The multi-class measure problem in nearest neighbour discrimination rules. Pattern Recognition, 23:1291?1297, 1990. [10] D. Ormoneit and T. Hastie. Optimal kernel shapes for local linear regression. In S. A. Solla, T. K. Leen, and K-R. Mller, editors, Advances in Neural Information Processing Systems, volume 12. MIT Press, 2000. [11] Sam Roweis and Lawrence Saul. Nonlinear dimensionality reduction by locally linear embedding. Science, 290(5500):2323?2326, Dec. 2000. [12] R. D. Short and K. Fukunaga. The optimal distance measure for nearest neighbor classification. IEEE Transactions on Information Theory, 27:622?627, 1981. [13] P. Y. Simard, Y. A. LeCun, J. S. Denker, and B. Victorri. Transformation invariance in pattern recognition ? tangent distance and tangent propagation. Lecture Notes in Computer Science, 1524, 1998. [14] S. Tong and D. Koller. Restricted bayes optimal classifiers. In Proceedings of the 17th National Conference on Artificial Intelligence (AAAI), pages 658?664, Austin, Texas, 2000. [15] V.N. Vapnik. The Nature of Statistical Learning Theory. Springer, New York, 1995. [16] Bin Zhang. Is the maximal margin hyperplane special in a feature space? Technical Report HPL-2001-89, Hewlett-Packards Labs, 2001. Table 1: Test-error obtained on the USPS and MNIST digit classification tasks by KNN, SVM (using a Gaussian Kernel), HKNN and CKNN. Hyper parameters were tuned on a separate validation set. Both HKNN and CKNN appear to perform much better than original KNN, and even compare favorably to SVMs. Data Set USPS (6291 train, 1000 valid., 2007 test points) MNIST (50000 train, 10000 valid., 10000 test points) Algorithm KNN SVM HKNN CKNN KNN SVM HKNN CKNN Test Error 4.98% 4.33% 3.93% 3.98% 2.95% 1.30% 1.26% 1.46% Parameters used l l l l l  &    & U(U  &f  U   U     &    & U'U U  U l  0.032 CKNN basic HKNN HKNN, lambda=1 HKNN, lambda=10 0.03 0.028 error rate 0.026 0.024 0.022 0.02 0.018 0.016 0.014 0.012 0 20 40 60 80 100 120 K Figure 3: Error rate on MNIST as a function of l for CKNN, and HKNN with different values of . As can be seen the basic HKNN algorithm performs poorly for large values of l . As expected, CKNN is relatively unaffected by this problem, and HKNN can be made robust through the added ?weight decay? penalty controlled by . Table 2: Test-error obtained on MNIST with HKNN and CKNN when using a reduced training set made of the 16712 support vectors retained by the best Gaussian Kernel SVM. This corresponds to 28% of the initial 60000 training patterns. Performance is even better than when using the whole dataset. But here, hyper parameters l and were chosen with the test set, as we did not have a separate validation set in this setting. It is nevertheless remarkable that comparable performances can be achieved with far fewer points. Data Set MNIST (16712 train s.v., 10000 test points) Algorithm HKNN CKNN Test Error 1.23% 1.36%  U  Parameters   used & l l  U
2058 |@word version:1 middle:1 seems:1 nonsensical:1 covariance:1 dramatic:1 mention:3 reduction:2 myles:1 initial:4 tuned:1 outperforms:1 yet:1 written:1 visible:1 partition:1 shape:1 wanted:1 remove:1 discrimination:2 intelligence:2 fewer:2 short:3 hyperplanes:1 zhang:1 along:1 direct:1 qualitative:1 prove:1 consists:1 combine:1 indeed:1 expected:2 frequently:1 multi:2 actual:1 encouraging:1 considering:2 estimating:1 linearity:1 maximizes:1 what:2 fantasy:1 substantially:1 vicinal:1 finding:1 transformation:3 remember:1 quantitative:1 nf:1 universit:1 classifier:2 appear:2 producing:1 safety:1 understood:1 local:23 limit:2 mistake:1 depended:1 might:1 twice:1 suggests:1 decided:2 unique:1 lecun:1 testing:1 practice:1 implement:2 digit:2 suggest:1 get:3 cannot:1 close:4 risk:1 www:1 equivalent:1 center:1 maximizing:2 missing:3 graphically:1 latest:1 lfl:1 convex:6 qc:1 simplicity:1 rule:1 embedding:2 notion:3 variation:1 hurt:1 exact:1 programming:1 olivier:1 hjilk:1 origin:1 trick:2 associate:1 approximated:1 recognition:3 solved:1 capture:1 region:4 solla:1 mentioned:3 intuition:6 mozer:1 broken:1 ideally:1 solving:5 usps:2 easily:1 train:3 fast:1 effective:1 artificial:2 hyper:2 neighborhood:7 choosing:1 quite:1 whose:5 stanford:1 otherwise:1 statistic:1 knn:17 unseen:1 think:1 h3c:1 final:1 advantage:1 eigenvalue:2 propose:1 reconstruction:2 maximal:5 adaptation:1 relevant:1 poorly:2 achieve:1 roweis:1 extending:2 produce:2 derive:1 montreal:1 fixing:2 measured:2 nearest:25 indicate:1 differ:1 direction:2 guided:1 drawback:1 correct:1 hull:4 vc:1 bin:1 fix:1 generalization:2 really:2 strictly:1 hold:1 around:1 considered:5 residing:1 cb:1 mapping:2 lawrence:1 mller:1 achieves:1 smallest:5 purpose:1 perceived:1 label:2 largest:2 hasselmo:1 weighted:1 minimization:1 mit:2 j7:1 gaussian:3 modified:5 rather:5 avoid:1 derived:1 schaal:1 improvement:1 logically:1 greatly:1 centroid:3 flaw:1 dependent:1 nn:6 typically:2 koller:1 interested:2 issue:1 classification:11 among:4 pascal:1 flexible:1 constrained:1 special:1 never:2 barring:1 having:2 look:3 yoshua:1 report:2 fundamentally:1 duplicate:1 few:1 neighbour:1 national:1 dfg:2 ourselves:2 attempt:1 freedom:1 friedman:1 montr:1 possibility:1 hewlett:1 closer:1 euclidean:4 old:2 penalizes:1 instance:2 eal:1 column:2 earlier:1 cover:1 introducing:1 subset:2 rare:1 too:3 nns:1 density:3 michael:2 quickly:1 again:1 aaai:1 possibly:1 worse:1 lambda:2 simard:1 toy:1 li:1 de:1 explicitly:1 later:1 try:2 view:2 performed:1 jason:1 lab:1 bayes:2 minimize:2 accuracy:1 largely:1 maximized:1 conceptually:1 generalize:1 vincent:1 produced:7 lu:1 straight:1 unaffected:1 touretzky:1 checked:1 trevor:1 definition:1 associated:2 di:2 dataset:1 knowledge:3 dimensionality:3 cj:1 actually:1 appears:2 higher:1 x6:1 improved:2 formulation:2 leen:2 hpl:1 hand:2 nonlinear:1 propagation:1 somehow:1 artifact:1 name:1 dietterich:1 building:4 concept:1 contain:1 requiring:1 hence:1 equality:1 moore:1 attractive:1 performs:1 geometrical:4 recently:1 umontreal:2 volume:3 belong:1 refer:1 fantasized:2 had:1 moving:1 chapelle:1 similarity:1 surface:26 closest:7 certain:1 success:1 seen:4 additional:2 lowerdimensional:1 somewhat:1 paradigm:2 smooth:1 technical:2 hart:1 controlled:1 impact:1 variant:1 basic:6 regression:2 expectation:1 metric:2 sometimes:1 kernel:7 achieved:1 dec:1 penalize:1 whereas:1 want:1 fine:1 victorri:1 unlike:1 pass:1 induced:2 virtually:1 member:1 spirit:2 seem:1 call:1 near:1 abusively:1 ideal:1 bengio:1 easy:1 affect:2 hastie:2 restrict:3 idea:4 prototype:1 avenue:1 texas:1 nfl:2 whether:1 penalty:3 suffer:1 speaking:1 york:1 useful:1 clear:1 informally:1 amount:1 locally:10 svms:13 reduced:1 http:1 generate:1 notice:1 arising:2 correctly:3 tibshirani:1 shall:1 nevertheless:1 falling:1 drawn:2 angle:1 everywhere:1 almost:1 reasonable:1 strange:1 decide:1 guyon:1 decision:26 comparable:1 quadratic:1 annual:1 fading:1 precisely:1 constraint:3 dominated:1 speed:1 fukunaga:1 performing:1 separable:3 relatively:3 department:1 popularized:1 combination:3 disconnected:2 poor:2 smaller:2 sam:1 appealing:1 intuitively:1 restricted:1 jlk:1 equation:3 remains:1 previously:1 available:1 denker:1 ocr:1 away:3 appearing:1 original:4 denotes:2 top:1 eon:1 build:1 especially:2 classical:3 objective:1 question:1 added:2 usual:1 diagonal:1 subspace:5 distance:23 separate:3 link:1 separating:2 capacity:2 manifold:10 considers:1 discriminant:1 iro:3 besides:1 retained:1 illustration:1 vladimir:1 mostly:2 robert:1 favorably:1 cknn:15 ba:2 unknown:1 perform:5 allowing:1 finite:2 looking:1 rections:2 canada:1 inferred:2 david:1 complement:1 pair:2 namely:1 required:1 boser:1 boost:1 able:1 usually:1 pattern:7 vincentp:2 including:1 memory:2 explanation:1 natural:1 indicator:1 ormoneit:1 representing:1 kclass:1 tresp:1 prior:3 nice:1 geometric:1 tangent:6 understanding:1 review:1 lecture:1 highlight:1 interesting:4 remarkable:1 validation:2 degree:1 affine:5 editor:3 share:1 austin:1 surprisingly:1 keeping:1 soon:1 allow:1 understand:1 neighbor:29 saul:1 face:1 fifth:1 regard:1 overcome:2 plain:1 dimension:3 world:2 boundary:2 rich:1 contour:1 valid:3 doesn:1 made:3 adaptive:1 atkeson:1 far:2 transaction:3 implicitly:1 keep:2 dealing:1 rid:1 pittsburgh:1 assumed:1 alternatively:2 search:2 why:3 table:4 promising:1 nature:2 reasonably:1 robust:1 ca:2 ignoring:1 bottou:2 complex:3 tween:1 did:2 main:1 linearly:2 motivation:1 noise:2 arise:1 whole:3 enriched:1 referred:1 representative:1 rithm:1 slow:1 tong:1 bengioy:1 third:1 jmk:1 decay:3 svm:14 incorporating:1 essential:2 mnist:6 vapnik:4 adding:1 workshop:1 illustrates:3 hole:2 margin:16 likely:2 explore:1 happening:1 expressed:2 partially:1 springer:1 corresponds:4 weston:1 conditional:1 identity:2 loosing:1 hard:1 infinite:3 typical:2 reducing:1 hyperplane:19 principal:1 called:1 invariance:5 experimental:2 meaningful:1 formally:2 support:7 arises:1 absolutely:1 dept:1 handling:1
1,161
2,059
Kernel Logistic Regression and the Import Vector Machine Trevor Hastie Department of Statistics Stanford University Stanford, CA 94305 hastie@stat.stanford.edu Ji Zhu Department of Statistics Stanford University Stanford, CA 94305 jzhu@stat.stanford.edu Abstract The support vector machine (SVM) is known for its good performance in binary classification, but its extension to multi-class classification is still an on-going research issue. In this paper, we propose a new approach for classification, called the import vector machine (IVM), which is built on kernel logistic regression (KLR). We show that the IVM not only performs as well as the SVM in binary classification, but also can naturally be generalized to the multi-class case. Furthermore, the IVM provides an estimate of the underlying probability. Similar to the ?support points? of the SVM, the IVM model uses only a fraction of the training data to index kernel basis functions, typically a much smaller fraction than the SVM. This gives the IVM a computational advantage over the SVM, especially when the size of the training data set is large. 1 Introduction      In standard classification problems, we are given a set of training data , ,    , where the output is qualitative and assumes values in a finite set  . We  wish to find a classfication rule from the training data, so that when given a new input , we can assign a class  from  to it. Usually it is assumed that the training data are an independently and identically distributed sample from an unknown probability distribution   . "!$#&%'() The support vector machine (SVM) works well in binary classification, i.e. , but its appropriate extension to the multi-class case is stillan on-going research issue. Another 3$(54687 9 weakness of the SVM is that it only estimates , while the probability 2 *+-,/.10 2 9 :  ;:<(/= >:? is often of interest itself,( where 2 @:A is the conditional probability of a point being in class given . In this paper, we propose a new approach, called the import vector machine (IVM), to address the classification problem. We show that the IVM not only performs as well as the SVM in binary classification, but also can naturally be generalized to the multi-class case. Furthermore, the IVM provides an estimate of the  probability 2 . Similar to the ?support points? of the SVM, the IVM model uses only a fraction of the training data to index the kernel basis functions. We call these training data C$D& &E8 E cost of the SVM is B E import points. The computational , while the computational cost C of the IVM is B , where is the number of import points. Since does not tend to C increase as increases, the IVM can be faster than the SVM, especially for large training data sets. Empirical results show that the number of import points is usually much less than the number of support points. In section (2), we briefly review some results of the SVM for binary classification and compare it with kernel logistic regression (KLR). In section (3), we propose our IVM algorithm. In section (4), we show some simulation results. In section (5), we generalize the IVM to the multi-class case. 2 Support vector machines and kernel logistic regression The standard SVM produces a non-linear classification boundary in the original input space by constructing a linear boundary in a transformed version of the original input space. The dimension of the transformed space can be very large, even infinite in some cases. This seemingly prohibitive computation is achieved through a positive definite reproducing kernel , which gives the inner product in the transformed space. Many people have noted the relationship between the SVM and regularized function estimation in the reproducing kernel Hilbert spaces (RKHS). An overview can be found in Evgeniou et al. (1999), Hastie et al. (2001) and Wahba (1998). Fitting an SVM is equivalent to minimizing:  ( (1)  C :  !  ( 3$           ! with .  7 classification rule is given by *+-,/.10 .    is the RKHS generated by the kernel  By the representer theorem (Kimeldorf et al  (1971)), the optimal (2)    :      C 9 . The has the form:   It often happens that a sizeable fraction of the values of  can be zero. This is a consequence of the truncation property of the first part of criterion (1). This seems to be an attractive property, because only the points on the wrong side of the classification boundary, and those on the right side but near the boundary have an influence in determining the 9  position of the boundary, and hence have non-zero  ?s. The corresponding ?s are called support points.   ( 3  & Notice that (1) has the form !*&* 2#"&.  %$ . The loss function is plotted in Figure 1, along with several traditional loss functions. As we can see, the negative log-likelihood (NLL) of the binomial distribution has a similar shape to that of the SVM. If we replace ( 3   ' (* ",+.-0/ in (1) with () , the NLL of the binomial distribution, the problem becomes a KLR problem. We expect that the fitted function performs similarly to the SVM for binary classfication. There are two immediate advantages of making such a replacement: (a) Besides giving 9 : a classification rule, the KLR also offers a natural estimate of the probability 2 4 (*  3 (54687 "1/ "1/ , while the SVM only estimates *+-, .10 2 ; (b) The KLR can naturally be generalized to the multi-class case through kernel multi-logit regression, whereas this is not the case for the SVM. However, because the KLR compromises the hinge loss function of the SVM, it no longer has the ?support points? property; in other words, all the  ?s in (2) are non-zero.  KLR is a well studied problem; see Wahba et al. (1995) and references there; see also Green et al. (1985) and Hastie et al. (1990). 3.0 0.0 0.5 1.0 1.5 Loss 2.0 2.5 Binomial NLL Squared Error Support Vector -3 -2 -1 0 1 2 3 yf(x) Figure 1: Several loss functions, C   D5 The computational cost of the KLR is B ; to save the computational cost, the IVM algorithm will find a sub-model to approximate the full model (2) given by the KLR. The sub-model has the form:  1:  (3)          # 8   )  where is a subset of the training data , and the data in are called import points. The advantage of this sub-model is that the computational cost is reduced, especially for large training data sets, while not jeopardizing the performance in classification.  Several other researchers have investigated techniques in selecting the subset . Lin et al. (1998) divide the training data into several clusters, then randomly select a representative from each clusterE to make up . Smola et al. (2000)  develope a greedy technique to se  & 7  E 0 , such that the span of quentially select columns of the kernel matrix   7   E these columns approximates the span of 0 well in the Frobenius norm. Williams et al. (2001) propose randomly selecting points of the training data, then using the Nystrom method to approximate the eigen-decomposition of the kernel matrix    & 7  C  0 , and expanding the results back up to dimensions. None of these meth   ods uses the output in selecting the/subset (i.e., the procedure only involves ). The   IVM algorithm uses both the output and the input to select the subset , in such a way that the resulting fit approximates the full model well.          3 Import vector machine / ! #5% () Following the tradition of logistic regression, we let for the rest of this paper. For notational simplicity, the constant term in the fitted function is ignored. In the KLR, we want to minimize:   :  3    0    3  ( ( )       7   6   From (2), it can be shown that this is equivalent to the finite dimensional form: (4)  : ! #"  3    $ ( () ( % ! #"      6   '&    : where  tion matrix  8 '& :    '" ; the regressor matrix #" .  :    0 7   ; and the regulariza-   To find  , we set the derivative of with respect to  equal to 0, and use the NewtonRaphson method to iteratively solve the score equation. It can be shown that the NewtonRaphson step is a weighted least squares step:  (5)      " :    #" #& '"    :  where  is: the valueof ( in3 the  - th 7 step,  matrix is  . +  ,90 2 2 " +          +     + 3 #& 8   2   . The weight )  As mentioned in section 2, we want to find a subset of , such that the sub-model (3) is a good approximation of the full model (2). Since it is impossible to search for every subset , we use the following greedy forward strategy:  3.1 Basic algorithm ( ( ( 6  : ) Let ,  ) For each :  #&98  !    to minimize    (6)   ( '& .     3 : 3  :    0      # "     3  "  ( :  #  ) Let  ) Repeat steps ( We call the points in  6  :$#  , , ) and ( ' ) until     "   &  & ( ()       7         7  :     0    "! :     6    ,   &      &  !   7    6   # % )      :  ( ( ) $     where the regressor matrix 0  ! #  ) E ; the regularization matrix  # ) : = =  ; . ) Let   : argmin  (  (  :   Find : , , let   9  )   #&98      &    , )     !  ,  : ,  ( . converges. import points. 3.2 Revised algorithm 6  The above algorithm is computationally feasible, but in step ( ) we need to useE the Newton-Raphson method to find  iteratively. When the number of import points becomes large, the Newton-Raphson computation can be expensive. To reduce this computation, we use a further approximation.   Instead of iteratively computing  until it converges, we can just do a one-step iteration, and use it as an approximation to the converged one. To get a good approximation, we E take advantage of the fitted result from the current ?optimal? , i.e., the sub-model when = =: , and use it as the initial value. This one-step update is similar to the score test in E generalized linear models (GLM); but the latter does not have a penalty term. The updating C formula allows the weighted regression (5) to be computed in B time.   Hence, we have the revised step ( 6 ) for the basic algorithm: 6 (   " ! &  ) For each , correspondingly augment with a column, and with a column and a row. Use the updating formula to find  in (5). Compute (6). 3.3 Stopping rule for adding point to      In step ( '& ) of the basic algorithm, we need to decide whether has converged. A 8   be the sequence natural stopping rule is to look at the regularized NLL. Let   , of regularized NLL?s obtained in step ( & ). At each step , we compare   with   + : (   +  where is a pre-chosen small integer, for example . If the ratio is less : % %%'(  than some pre-chosen small number  , for example,  , we stop adding new import points to .    3.4 Choosing the regularization paramter So far, we have assumed that the regularization parameter is fixed. In practice, we also need to choose an ?optimal? . We can randomly split all the data into a training set and a tuning set, and use the misclassification error on the tuning set as a criterion for choosing . To reduce the computation, we take advantage of the fact that the regularized NLL converges faster for a larger . Thus, instead of running the entire revised algorithm for procedure, which combines both adding import points to each , we propose the following and choosing the optimal :  ( ( 6 ( ( ( &  ) Start with a large regularization parameter . ) Let  : , :  6 #&       ) , : ( . ) Run steps ( ), (  ) and ( & ) of the revised algorithm, until the stopping cri: #       ) . Along the way, also compute the  terion is satisfied at misclassfication error on the tuning set. ) Decrease  & to a smaller value. (  ) Repeat steps (   ) and (  & ), starting with  :A#&       & ) . We choose the optimal as the one that corresponds to the minimum misclassification error on the tuning set. 4 Simulation In this section, we use a simulation to illustrate the IVM method. The data in each class are generated from a mixture of Gaussians (Hastie et al. (2001)). The simulation results are shown in Figure 2. 4.1 Remarks The support points of the SVM are those which are close to the classification boundary or 9 ( 3   misclassified and usually have large weights [2 ]. The import points of the 2 IVM are those that decrease the regularized NLL the most, and can be either close to or far from the classification boundary. This 9difference is natural, because the SVM is only 3 (84687 concerned with the classification , while the IVM also focuses on the *+ , .10 2  unknown probability 2 . Though points away from the classification boundary do not contribute to determining the position ofthe classification boundary, they may contribute to estimating the unknown probability 2 . Figure 3 shows a comparison of the SVM and C$D& &E8 E the IVM. The total computational cost ofE the SVM is B , while the computational cost C of the IVM method is B , where is the number of import points. Since does not 240 Misclassification rate for different lambda?s Regularized NLL for the optimal lambda ? ? ? 0.34 250 Regularized NLL for different lambda?s ??????? ? ? 0 220 ??? ?? ????????????????????????????????????????? ??????????????????????? ? ?? 0.26 ?? ?? ? ?? ?? 50 100 150 200 No. of import points added 0 ? ? ? ? ? ? ??? ?? ??? ?????? ?????????????????????????????????????????????????????????????????????????????? 50 100 150 200 No. of import points added                !"#  Figure 2: Radial kernel is used. , , the optimal . rate is found to correspond to stopping criterion is satisfied when ? 180 0.24 ?? ? ?? ???? ? ??? ?? ? ?? ?? ???????? 0.22 100 150 ? ???? ??? ?? ???? ?? ?? ?? ????????????? ?????????????????????????????????????????????????? 200 0.28 ?? ? ? ? 160 0.32 ? 200 ???? 0.30 ? 0 50 100 150 200 No. of import points added   . The left and middle panels illustrate how to choose decreases from to . The minimum misclassification . The right panel is for the optimal . The .   C tend to increase as increases, the computational cost of the IVM can be smaller than that of the SVM, especially for large training data sets. 5 Multi-class case $ &%(' &%*) $ $  In this section, we briefly describe a generalization of the IVM to multi-class classification.  (  Suppose there are classes. We can write the response as an -vector , with  each :  component being either 0 or 1, indicating which class the observation is in.  Therefore (  : %  :  : %' indicates the response is in the th class, and A( ( indicates the response is in the th 4 class. Using  the th class as the basis, the     : % : 4    :  ( ) 2 2 ( ) 2 2 ,  , . multi-logit can be written as Hence the Bayes classification rule is given by:  $ .- : %  argmax  -$ 0 1032 3240        (7) where  : :  3    0   5          - 3 ( ()  ,     " /  1:  :    -       " /  8           : + 65 $ 87977 6:  - /- .- We use + to index the observations, to index the classes, i.e. Then the regularized negative log-likelihood is (  7  -  6 -  C  % ,   :      %  (  $ . , and    :  Using the representer theorem (Kimeldorf et al. (1971)), the th element of which minimizes has the form  (8) +%,) $       ,    , SVM - with 107 support points IVM - with 21 import points . . . . . . . . . . . . . . +++ ++++++++++++++++++++++++++++++++++++++++++++++++++++ ++++++++++++++++++++++++++++++++++++++++++++++++++++ .. ... ... ... ... ... ... ... ... ... ... ... ... ... ++++ +++++++++++++++++++++++++++++++++++++++++++++++++++ ++++ +++++++++++++++++++++++++++++++++++++++++++++++++++ ++++ . . . . . . . . . . . . .. +++++ +++++++++++++++++++++++++++++++++++++++++++++++++++ ++++++++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++ +++++ ++++++++++++++++++++++++++++++++++++++++++++++++++ ++++++++++++++++++++++++++++++++++++++++++++++++++ . ... ... ... ... ... ... ... ... ... ... ... ... .. ++++++ +++++++++++++++++++++++++++++++++++++++++++++++++ ++++++ . +++++++++++++++++++++++++++++++++++++++++++++++++ +++++++ +++++++++++++++++++++++++++++++++++++++++++++++++ .. ... ... ... ... ... ... ... ... ... ... ... ... .. +++++++ ++++++++++++++++++++++++++++++++++++++++++++++++ +++++++ ++++++++++++++++++++++++++++++++++++++++++++++++ ++++++++ ++++++++++++++++++++++++++++++++++++++++++++++++ .. ... ... ... ... ... ... ... ... ... ... ... ... ... ++++++++ +++++++++++++++++++++++++++++++++++++++++++++++ ++++++++ +++++++++++++++++++++++++++++++++++++++++++++++ . . . . . . . . . . . . . ++++++++ +++++++++++++++++++++++++++++++++++++++++++++++ +++++++++++++++++++++++++++++++++++++++++++++++. .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++++ +++++++++ ++++++++++++++++++++++++++++++++++++++++++++++ . . . . . . . . . . . . . . +++++++++ ++++++++++++++++++++++++++++++++++++++++++++++ +++++++++ ++++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ ++++++++++++++++++++++++++++++++++++++++++++++ ++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ ... ... ... ... ... ... ... ... ... ... ... ... ... ... ++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ ++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ ++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ . . . . . . . . . . . . . . ++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ ++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ +++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++++++ +++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. +++++++++++ ++++++++++++++++++++++++++++++++++++++++++++++ +++++++++++ ++++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. +++++++++++ ++++++++++++++++++++++++++++++++++++++++++++++ +++++++++++ ++++++++++++++++++++++++++++++++++++++++++++++ . . . . . . . . . . . . +++++++++++ ++++++++++++++++++++++++++++++++++++++++++++++ ..... . . . . . . . . . . . +++++++++++ . +++++++++++++++++++++++++++++ +++++++++++++++++++++++++++. .. .. .. .. .. .. ++++++++++++++ +++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ ++++++++++ ++++++++++++++++++++++++++ . . . . . . . . .. +++++++++++++ . ++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ +++++++++++++++++++++++++ +++++++++ .. +++++++++++++++++++++++++ . .. .. .. .. .. .. .. .. .. .. ++++++++++++ . ++++++++++++ .. .. .. .. .. .. .. .. .. .. .. +++++++++ ++++++++++++++++++++++++ +++++++++ .. +++++++++++++++++++++++ . ... ... ... ... ... ... ... ... ... ... ... ... ... ++++++++++++ . .. .. .. .. .. .. .. .. .. .. +++++++++ . +++++++ .. .. ++++++++++++++++++++++ ++++++++++++ . . ++++++++++++++++++++++ .. +++++++++++ .. .. .. .. .. .. .. .. .. .. .. ... +++++++ +++++++++++++++++++++. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++++ +++++++. .. .. ++++++++++++++++++++ +++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. ++++++ .. +++++. .. .. .. +++++++++++++++++++ +++++++++++++++++++ +++++++++++ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ++++++++++++++++++. . . . . . . . . . . . . . . . . . . . . ++++++++++ ++++++++++ . . . . . . . . . . . .. .. ++++ ++++ . ++ . .. .. .. .. +++++++++++++++++ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ++++++++++++++++ ++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++++. .. .. .. .. .. .. .. .. .. .. .. ... ... ... ... ... ... ... ... ... ... +++++++++++++++ +++++++++++++++ ++++++++ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ++++++++++++++ .. +++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. ... ... ... ... ... ... ... ... ... +++++++++++++. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++ . . ++++++++++++ +++ ++++++++++++. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++++++ ++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ ............................................................ +++++++++ +++++++++. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++ ++++++++. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++ .............................................................. +++++++ ++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++ . . . . +++++ . . . . .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++++ ................................................................. ++++ ++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++ ++++. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++ .................................................................. +++ +++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++ . +++ . .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++ .................................................................. +++ +++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++ +++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++ .. .. .. .. .. .. Error: . . . . . . 0.160 ..................................................... Training ++++ ++++ . . . . . . .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++ ++++ . .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++++ .. .. .. .. .. .. .. .. .. .. 0.218 ..................................................... Test. Error: +++++ ++++++ . .. .. .. .. .. .. .. .. .. ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ++++++ +++++++ . . . . . . . . 0.210 ..................................................... +++++++ Bayes. Error: .. .. .. .. .. .. .. .. .. .. .. .. +++++++++++ ++++++++++++++++++++++++++++++++++++++++++++++ ++++++++++++++++++++++++++++++++++++++++++++++ +++++++++++ ++++++++++++++++++++++++++++++++++++++++++++++ . .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ ++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ ++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ . .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ ++++++++++ . ++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ ++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ ++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ ++++++++++ . . . . . . . . . . . . . . . ++++++++++++++++++++++++++++++++++++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ ++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ ++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ . . . . . . . . . . . . . . . ++++++++++ ++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ ++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++++++ +++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ +++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++++++ +++++++++++ ++++++++++++++++++++++++++++++++++++++++++++ . .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++++++ . ++++++++++++++++++++++++++++++++++++++++++++ ++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ . . . . . . . . . . . . . ++++++++++ ... +++++++++++++++++++++++++++++++++++++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ ++++++++++ +++++++++++++++++++++++++++++++++++++++++++++ . . . . . . . . . . . . . ++++++++++ ... +++++++++++++++++++++++++++++++++++++++++++++ ++++++++++++++++++++++++++++++++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ ++++++++++ .. +++++++++. .. ++++++++++++++++++++++++++++++++++++++++++++++ . . . . . . . . . . . . ++++++++++++++++++++++++++++++++++++++++++++++ ++++++++ . . +++++++++++++++++++++++++++++ . . .. ++++++++++++++++ . . +++++++++++++++. ... ... ... ... ... ... ... ... ... ... ... ... ++++++++ +++++++++++++++++++++++++++ ++++++++ .. .. ++++++++++++++++++++++++++ . .. .. .. .. .. .. .. ++++++++++++++ . +++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. ++++++++ +++++++++++++++++++++++++ ++++++++ .. .. +++++++++++++++++++++++++ .. ... ... ... ... ... ... ... ... ... +++++++++++++ .. ++++++++++++ . .. .. .. .. .. .. .. .. .. .. .. ++++++++ .. .. ++++++++++++++++++++++++ ++++++++++++++++++++++++ +++++++++++++ . . . . . . . . . . . ++++++++ ++++++++ +++++++++++++++++++++++ +++++++ .. .. .. .. ... ... ... ... ... ... ... ... ... ... ... ... +++++++++++++ .. ++++++++++++ .. .. .. .. .. .. .. .. .. .. .. +++++++ ++++++++++++++++++++++ ... ++++++++++++++++++++++ . .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++++ . .. .. .. .. .. .. .. .. .. .. ... ++++++ +++++++++++++++++++++ ++++++++++++ ++++++ .. .. .. . . +++++++++++++++++++++ ++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. ++++++ .. .. .. ++++++++++++++++++++ ++++++ .. ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... +++++++++++ +++++++++++++++++++ +++++++++++ ++++++ . . . . . . . . . . . . . . . +++++++++++++++++++ +++++++++++ ++++ . . .. .. ++++++++++++++++++ . .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ ... ... ... ... ... ... ... ... ... ... ... ... ... ++++ .. .. .. .. +++++++++++++++++ ++++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ ++++++++++. . . . . . . . . . . . . . ++++ .. +++ ++++++++++++++++ ++ . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. ++++++++ +++++++++++++++ .. ++++++ .. ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ++++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++ . . . . ++++++++++++++ ++ +++++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++++++++ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ++++++++++++ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .. .. .. ++++++++++++ .......................................................... +++++++++++ +++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. +++++++++++ ++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++++++++++ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .+ +++++++++ +++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++ +++++++++ ++ . . . ++++++++ .. .. .. ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ++ ++++++++ ++ ++++++++ ++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++ ++ ++++++++ ++++++++. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++ ++ +++++++ ++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++ +++++++ +++++++ ++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ++ ++++++++ ++++++++ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ++ ++++++++ + .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. ... + ++++++++ ++++++++ . .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. + ++++++++ +++++++++ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . + + +++++++++ .. .. .. .. .. .. 0.150 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .+ Training.Error: +++++++++ + ++++++++++ + . .. .. .. .. .. ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... + ++++++++++ +++++++++++ .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. .. + ++++++++++++ + Test Error: .. ... ... ... 0.219 ++++++++++++ + +++++++++++++ + . .. ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... ... + ++++++++++++++ +++++++++++++++ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .+ ++++++++++++++++ + Bayes Error: 0.210 ?? ? ? ? o ?o ? ? ? ???? ? ? ? ?? ? ?? o ? ? ? o o o ? ? ?o ??? ? ? ?? ??o o ? ?? o o ?o? ?? ? ? ? o ? oo o ?o?????? o ?? ? ???? o oooo ? o ooo o o ?? ? ? oo ? ? o?o ? ? ? ?o oooo o ????? ?? o o ? o o o oo? o ??? o ? oo oooo o ooo oo? ? o ? ? o?oo? oo o ? ? ?? o ?? ? ? ? o ?o ? ? ? ???? ? ? ? ?? ? ?? o ? ? ? o o o ? ? ?o ??? ? ? ?? ??o o ? ?? o o ?o? ?? ? ? ? o ? oo o ?o?????? o ?? ? ???? o oooo ? o ooo o o ?? ? ? oo ? ? o?o ? ? ? ?o oooo o ????? ?? o o ? o o o oo? o ??? o ? oo oooo o ooo oo? ? o ? ? o?oo? oo o o o o o o o o o oo oo o o ooo o o o oo o o oo o o o o o o o o o o o o o oo oo o o ooo o o o oo o o oo o o o o o ? ? ?? o Figure 3: The solid lines are the classification boundaries; the dotted lines are the Bayes rule (    boundaries. For the SVM, the dashed lines are the edges of the margin. For the IVM, the dashed lines are the  and  lines. Hence, (7) becomes   (9)  where case; and $ : : "  3     0   +    -  "     : +    , '" 3 is the + th row of () and ". ( '& (   "  0    7  6       &  are defined in the same way as in the binary E8 The multi-class IVM procedure is similar to the binary case, and the computational cost is C . Figure 4 is a simulation of the multi-class IVM. The data in each class are B generated from a mixture of Gaussians (Hastie et al. (2001)). Multi-class IVM - with 32 import points xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxx + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xx ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx x + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + ++++++xxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + ++ +++++++++xxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + ++ ++ ++ xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + ++ xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx ++ + + ++ ++ ++ + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + ++ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + ++ ++ ++ ++ ++ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx ++ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + ++ xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + ++ ++ ++ ++ ++ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +.......................................... ++...... xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx xxxxxxxxxxxxxxxxxxx............... ++ ++ ++ ++ ++ + + + + + + + + + + + + + + + + + + + + + + + + + + + + +........................................... xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx ....... + + + + + + + + + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx ..............+++++++++++++++++++++++............................................. ............................................ xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx ................... + + + + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxxxxxxxx .......................++++++++++++++++++............................................... .............................................. xxxxxxxxxxxxxxxxxxxxxxxxxxx .......................... + + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxxxxxxxxx .............................+++++++++++++++++................................................ ............................................... xxxxxxxxxxxxxxxxxxxxx ................................ + + + + + + + + + + + + + + + + xxxxxxxxxxxxxxxxx .................................... ................................................ ++ + + + + + + + + + + + + + + + + +................................................. xxxxxxxxxxxx ........................................... + + + + + + + + + + + + + + + + + xxxx .............................................. ................................................. + ++ + + + + + + + + + + + + + + + + +.................................................. .............................................. + + + + + + + + + + + + + + + + + .............................................. .................................................. ++ + + + + + + + + + + + + + + + + + ............................................... .................................................... +++ + ................................................. ++ ++ ++ ++ ++ ++ ++ ++ ++ ++ ++...................................................... .................................................................................................................. .................................................................................................................. .................................................................................................................. .................................................................................................................. .................................................................................................................. .................................................................................................................. .................................................................................................................. .................................................................................................................. .................................................................................................................. .................................................................................................................. .................................................................................................................. .................................................................................................................. .................................................................................................................. .................................................................................................................. Training Error: 0.237 .................................................................................................................. .................................................................................................................. .................................................................................................................. Test Error: 0.259 .................................................................................................................. .................................................................................................................. .................................................................................................................. .................................................................................................................. o o o o ooo oo oo o o o oo o oo o ooooooo o o o o o o o oooo o o ooo o ooo o o o o oo o oo o oo o oo o o o ooo o o o o o oo oo o oo o o oo o o o o o oo o o o o oo o Bayes Error: 0.251 Figure 4: Radial kernel is used.   ( (  (  ! # , , , . 6 Conclusion We have discussed the import vector machine (IVM) method in both binary and multi-class E8 classification. We showed that it 9not only performs as well as the SVM, but alsoC provides E8 2 E the IVM is B an estimate of the probability . The computational cost of for C the binary case and B for the multi-class case, where is the number of import points. $ Acknowledgments We thank Dylan Small, John Storey, Rob Tibshirani, and Jingming Yan for their helpful comments. Ji Zhu is partially supported by the Stanford Graduate Fellowship. Trevor Hastie is partially supported by grant DMS-9803645 from the National Science Foundation, and grant ROI-CA-72028-01 from the National Institutes of Health. Thanks to Grace Wahba and Chris Williams for pointing out several interesting and important references. We also want to thank the anonymous NIPS referees who helped improve this paper. References [1] Burges, C.J.C. (1998) A tutorial on support vector machines for pattern recognition. In Data Mining and Knowledge Discovery. Kluwer Academic Publishers, Boston. (Volume 2) [2] Evgeniou, T., Pontil, M., & Poggio., T. (1999) Regularization networks and support vector machines. In A.J. Smola, P. Bartlett, B. Sch?olkopf, and C. Schuurmans, editors, Advances in Large Margin Classifiers. MIT Press. [3] Green, P. & Yandell, B. (1985) Semi-parametric generalized linear models. Proceedings 2nd International GLIM Conference, Lancaster, Lecture notes in Statistics No. 32 44-55 Springer-Verlag, New York. [4] Hastie, T. & Tibshirani, R. (1990) Generalized Additive Models, Chapman and Hall. [5] Hastie, T., Tibshirani, R., & Friedman, J.(2001) The elements of statistical learning. In print. [6] Lin, X., Wahba, G., Xiang, D., Gao, F., Klein, R. & Klein B. (1998), Smoothing spline ANOVA models for large data sets with Bernoulli observations and the randomized GACV. Technical Report 998, Department of Statistics, University of Wisconsin, Madison WI. [7] Kimeldorf, G. & Wahba, G. (1971) Some results on Tchebycheffian spline functions. J. Math. Anal. Applic. 33, 82-95. [8] Smola, A. & Sch?olkopf, B. (2000) Sparse Greedy Matrix Approximation for Machine Learning. In Proceedings of the Seventeenth International Conference on Machine Learning. Morgan Kaufmann Publishers. [9] Wahba, G. (1998) Support Vector Machine, Reproducing Kernel Hilbert Spaces and the Randomized GACV. Technical Report 984rr, Department of Statistics, University of Wisconsin, Madison WI. [10] Wahba, G., Gu, C., Wang, Y., & Chappell, R. (1995) Soft Classification, a.k.a. Risk Estimation, via Penalized Log Likelihood and Smoothing Spline Analysis of Variance. In D.H. Wolpert, editor, The Mathematics of Generalization. Santa Fe Institute Studies in the Sciences of Complexity. Addison-Wesley Publisher. [11] Williams, C. & Seeger, M (2001) Using the Nystrom Method to Speed Up Kernel Machines. In T. K. Leen, T. G. Diettrich, and V. Tresp, editors, Advances in Neural Information Processing Systems 13. MIT Press.
2059 |@word version:1 briefly:2 middle:1 seems:1 norm:1 logit:2 nd:1 simulation:5 decomposition:1 usee:1 solid:1 initial:1 score:2 selecting:3 rkhs:2 terion:1 current:1 od:1 import:21 written:1 john:1 additive:1 shape:1 update:1 greedy:3 prohibitive:1 provides:3 math:1 contribute:2 along:2 qualitative:1 fitting:1 combine:1 multi:15 becomes:3 estimating:1 underlying:1 xx:1 panel:2 kimeldorf:3 argmin:1 minimizes:1 every:1 wrong:1 classifier:1 grant:2 positive:1 consequence:1 diettrich:1 studied:1 graduate:1 seventeenth:1 acknowledgment:1 practice:1 definite:1 ofe:1 procedure:3 pontil:1 empirical:1 yan:1 word:1 pre:2 radial:2 get:1 close:2 risk:1 influence:1 impossible:1 yandell:1 equivalent:2 williams:3 starting:1 independently:1 simplicity:1 rule:7 d5:1 suppose:1 us:4 element:2 referee:1 expensive:1 recognition:1 updating:2 wang:1 decrease:3 e8:5 mentioned:1 complexity:1 compromise:1 basis:3 gacv:2 gu:1 describe:1 choosing:3 lancaster:1 stanford:7 solve:1 larger:1 statistic:5 itself:1 seemingly:1 nll:9 advantage:5 sequence:1 rr:1 propose:5 product:1 frobenius:1 olkopf:2 cluster:1 produce:1 converges:3 illustrate:2 oo:36 stat:2 involves:1 assign:1 generalization:2 anonymous:1 extension:2 hall:1 roi:1 pointing:1 estimation:2 weighted:2 mit:2 focus:1 notational:1 bernoulli:1 likelihood:3 indicates:2 seeger:1 tradition:1 helpful:1 cri:1 stopping:4 typically:1 entire:1 going:2 transformed:3 misclassified:1 issue:2 classification:24 augment:1 smoothing:2 equal:1 evgeniou:2 chapman:1 look:1 representer:2 report:2 spline:3 randomly:3 national:2 argmax:1 replacement:1 friedman:1 interest:1 mining:1 weakness:1 mixture:2 edge:1 poggio:1 divide:1 plotted:1 fitted:3 column:4 soft:1 cost:10 subset:6 thanks:1 international:2 randomized:2 regressor:2 squared:1 satisfied:2 choose:3 lambda:3 derivative:1 sizeable:1 tion:1 helped:1 start:1 bayes:5 regulariza:1 minimize:2 square:1 kaufmann:1 who:1 variance:1 correspond:1 generalize:1 none:1 researcher:1 converged:2 trevor:2 nystrom:2 naturally:3 dm:1 stop:1 knowledge:1 oooo:7 hilbert:2 back:1 glim:1 wesley:1 response:3 ooo:10 leen:1 though:1 furthermore:2 just:1 smola:3 until:3 xxxx:2 logistic:5 yf:1 hence:4 regularization:5 iteratively:3 attractive:1 noted:1 criterion:3 generalized:6 performs:4 classfication:2 ji:2 overview:1 volume:1 discussed:1 approximates:2 kluwer:1 tuning:4 mathematics:1 similarly:1 longer:1 showed:1 verlag:1 binary:10 morgan:1 minimum:2 dashed:2 semi:1 full:3 technical:2 faster:2 academic:1 offer:1 raphson:2 lin:2 regression:7 basic:3 iteration:1 kernel:16 achieved:1 whereas:1 want:3 fellowship:1 publisher:3 sch:2 rest:1 comment:1 tend:2 call:2 integer:1 near:1 split:1 identically:1 concerned:1 fit:1 hastie:9 wahba:7 inner:1 reduce:2 whether:1 bartlett:1 penalty:1 york:1 remark:1 ignored:1 se:1 santa:1 reduced:1 tutorial:1 notice:1 dotted:1 tibshirani:3 klein:2 write:1 tchebycheffian:1 chappell:1 anova:1 fraction:4 run:1 decide:1 speed:1 span:2 department:4 smaller:3 wi:2 rob:1 making:1 happens:1 glm:1 computationally:1 equation:1 paramter:1 addison:1 gaussians:2 away:1 appropriate:1 save:1 eigen:1 original:2 binomial:3 assumes:1 running:1 hinge:1 newton:2 madison:2 giving:1 especially:4 added:3 print:1 strategy:1 parametric:1 traditional:1 grace:1 thank:2 chris:1 besides:1 index:4 relationship:1 ratio:1 minimizing:1 fe:1 negative:2 anal:1 unknown:3 observation:3 revised:4 finite:2 immediate:1 reproducing:3 applic:1 nip:1 address:1 in3:1 usually:3 pattern:1 built:1 green:2 misclassification:4 natural:3 regularized:8 zhu:2 meth:1 improve:1 health:1 tresp:1 review:1 discovery:1 determining:2 xiang:1 wisconsin:2 loss:5 expect:1 lecture:1 interesting:1 foundation:1 editor:3 klr:10 row:2 penalized:1 repeat:2 supported:2 truncation:1 side:2 burges:1 institute:2 correspondingly:1 sparse:1 distributed:1 boundary:11 dimension:2 forward:1 far:2 approximate:2 assumed:2 search:1 storey:1 ca:3 expanding:1 schuurmans:1 investigated:1 constructing:1 ooooooo:1 representative:1 sub:5 position:2 wish:1 dylan:1 theorem:2 formula:2 svm:28 adding:3 margin:2 boston:1 wolpert:1 gao:1 partially:2 springer:1 corresponds:1 ivm:29 conditional:1 replace:1 feasible:1 infinite:1 called:4 total:1 indicating:1 select:3 support:14 people:1 newtonraphson:2 latter:1
1,162
206
590 Atiya and Abu-Mostafa A Method for the Associative Storage of Analog Vectors Amir Atiya (*) and Yaser Abu-Mostafa (**) (*) Department of Electrical Engineering (**) Departments of Electrical Engineering and Computer Science California Institute Technology Pasadena, Ca 91125 ABSTRACT A method for storing analog vectors in Hopfield's continuous feedback model is proposed. By analog vectors we mean vectors whose components are real-valued. The vectors to be stored are set as equilibria of the network. The network model consists of one layer of visible neurons and one layer of hidden neurons. We propose a learning algorithm, which results in adjusting the positions of the equilibria, as well as guaranteeing their stability. Simulation results confirm the effectiveness of the method . 1 INTRODUCTION The associative storage of binary vectors using discrete feedback neural nets has been demonstrated by Hopfield (1982). This has attracted a lot of attention, and a number of alternative techniques using also the discrete feedback model have appeared. However, the problem of the distributed associative storage of analog vectors has received little attention in literature. By analog vectors we mean vectors whose components are real-valued. This problem is important because in a variety of applications of associative memories like pattern recognition and vector quantization the patterns are originally in analog form and therefore one can save having the costly quantization step and therefore also save increasing the dimension of the vectors. In dealing with analog vectors, we consider feedback networks of the continuous-time graded-output variety, e.g. Hopfield's model (1984): du dt = -u + Wf(u) + a, x = f(u), (1) where u = (Ul, ... , UN)T is the vector of neuron potentials, x = (x!, ... , XN)T is the vector of firing rates, W is the weight matrix, a is the threshold vector, and f(u) means the vector (f( uI), ... , f( UN)) T, where f is a sigmoid-shaped function. The vectors to be stored are set as equilibria of the network. Given a noisy version of any of the stored vectors as the initial state of the network, the network state has A Method for the Associative Storage of Analog Vectors to reach eventually the equilibrium state corresponding to the correct vector. An important requirement is that these equilibria be asymtotically stable, otherwise the attraction to the equilibria will not be guaranteed. Indeed, without enforcing this requirement, our numerical simulations show mostly unstable equilibria. 2 THE MODEL It can be shown that there are strong limitations on the set of memory vectors which can be stored using Hopfield's continuous model (Atiya and Abu-Mostafa 1990). To relieve these limitations, we use an architecture consisting of both visible and hidden units. The outputs of the visible units correspond to the components of the stored vector. Our proposed architecture will be close to the continuous version of the BAM (Kosko 1988). The model consists of one layer of visible units and another layer of hidden units (see Figure 1). The output of each layer is fed as an input to the other layer. No connections exist within each of the layers. Let y and x be the output vectors of the hidden layer and the visible layer respectively. Then, in our model, du dt = -u + Wf(z) + a = e, y = f(u) (2a) dz = -z + Vf(u) + b = h, x = f(z) = [Wij] and V = [Vij] are the weight matrices, a dt (2b) where W and b are the threshold vectors, and f is a sigmoid function (monotonically increasing) in the range from -1 to 1, for example f(u) = tanh(u). x x hld~n vlSlbl. l~y.,. l~y.,. Figure 1: The model 591 592 Atiya and Abu?Mostafa As we mentioned before, for a basin of attraction to exist around a given memory vector, the corresponding equilibrium has to be asymtotically stable. For the proposed architecture a condition for stability is given by the following theorem. Theorem: An equilibrium point (u*, z*) satisfying J'l/2( un 2:IWij If'l/2(zj) < 1 (3a) j J'l/\Z;) 2:I Vij l!,l/2(uj) < 1 (3b) j for all i is asymptotically stable. Proof: We linearize (2a), (2b) around the equilibrium. We get dq -=Jq, du where if i if i = 1, ... , Nl = Nl + 1, ..., Nl + N 2, Nl and N2 are the number of units in the hidden layer and the visible layer respectively, and J is the Jacobian matrix, given by ~ aUl J= ~ fu ~ ae~l ae~l ae~l aUl aUNl ah aZ 1 aeN1 aZN'J ~hl ~ ahNa ahNa aZN'J aUNl ghl Ul 8U";1 ahNa ahNa aUNl aUl aZ 1 Zl lhl aZN'J aZN'J the partial derivatives evaluated at the equilibrium point. Let Al and A2 be respectively the Nl x Nl and N2 x N2 diagonal matrices with the ith diagonal element being respectively f'(un and f'(z;). Furthermore, let The Jacobian is evaluated as where IL means the L x L identity matrix. Let ( A- _A- l 1 V A Method for the Associative Storage of Analog Vectors Then, J=AA. Eigenvalues of AA are identical to the eigenvalues of A 1/2 AA 1/2 because if ). is an eigenvalue of AA corresponding to eigenvector v, then AAv = ).v, and hence Now, we have Al/2AAI/2 _ (-INl A~/2V A~/2 A~/2WA~/2) -IN2 . By Gershgorin's Theorem (Franklin 1968), an eigenvalue of J has to satisfy at least one of the inequalities: I). + 11 ::; f'1/2( un 2:IWii 1f'1/2(zi) i = 1, ... ,N1 i I). + 11::; f'1/2(zn2:lvjil!,1/2(uj) i = 1, ... ,N2' i It follows that under conditions (3a), (3b) that the eigenvalues of J will have negative real parts, and hence the equilibrium of the original system (2a), (2b) will be asymptotically stable. Thus, if the hidden unit values are driven far enough into the saturation region (i.e . with values close to 1 or -1), then the corresponding equilibrium will be stable because then, 1'( will be very small, causing Inequalities (3) to be satisfied. Although there is nothing to rule out the existence of spurious equilibria and limit cycles, if they occur then they would be far away from the memory vectors because each memory vector has a basin of attraction around it. In our simulations we have never encountered limit cycles. un 3 TRAINING ALGORITHM = Let xm, m 1, ... , M be the vectors to be stored. Each xm should correspond to the visible layer component of one of the asymptotically stable equilibria. We design the network such that the hidden layer component of the equilibrium corresponding to xm is far into the saturation region. The target hidden layer component ym can be taken as a vector of l's and -1 's, chosen arbitrarily for example by generating the components randomly. Then, the weights have to satisfy yj = !(2:Wi/X, + aj), / xi = ![2:Vjj!(2:Wj/x/ j / + aj) + b;]. 593 594 Atiya and Abu-Mostafa Training is performed in two steps. In the first step we train the weights of the hidden layer. We use steepest descent on the error function El = Lllyj - f(LWjlX; + aj )11 2 . I m,j In the second step we train the weights of the visible layer, using steepest descent on the error function E2 =L II xi - ![LVij!(LWj/x; + aj) + bd 112. m,i j I We remark that in the first step convergence might be slow since the targets are lor -1. A way to have fast convergence is to stop if the outputs are within some constant (say 0.2) from the targets. Then we multiply the weights and the thresholds of the hidden layer by a big positive constant, so as to force the outputs of the hidden layer to be close to 1 or -1. 4 IMPLEMENTATION We consider a network with 10 visible and 10 hidden units. The memory vectors are randomly generated (the components are from -0.8 to 0.8 rather than the full range to have a faster convergence). Five memory vectors are considered. After learning, the memory is tested by giving memory vectors plus noise (100 vectors for a given variance). Figure 2 shows the percentage correct recall in terms of the signal to noise ratio. Although we found that we could store up to 10 vectors, working close to the full capacity is not recommended, as the recall accuracy dc>teriorates. /. correct 100 -r--.......--~~---------::::_-----> 80 60 40 20 O.f...o.-----------............-----I -6 6 10 -2 2 snr (db) Figure 2: Recall accuracy versus signal to noise ratio A Method for the Associative Storage of Analog Vectors Acknowledgement This work is supported by the Air Force Office of Scientific Research under grant AFO SR-88-0231 . References J. Hopfield (1982), "Neural networks and physical systems with emergent collective computational abilities", Proc. Nat. Acad. Sci. USA, vol. 79, pp. 2554-2558. J. Hopfield (1984), "Neurons with graded response have collective computational properties like those of two state neurons", Proc. Nat. Acad. Sci. USA, vol. 81, p. 3088-3092. A. Atiya and Y. Abu-Mostafa (1990), "An analog feedback associative memory", to be submitted. B. Kosko (1988), "Bidirectional associative memories", IEEE Trans. Syst. Man Cybern., vol. SMC-18, no. 1, pp. 49-60. J. Franklin (1968) Matrix Theory, Prentice-Hall, Englewood Cliffs, New Jersey. 595 PART VII: EMPIRICAL ANALYSES
206 |@word uj:2 graded:2 version:2 hence:2 correct:3 simulation:3 costly:1 diagonal:2 sci:2 capacity:1 initial:1 unstable:1 franklin:2 enforcing:1 around:3 considered:1 hall:1 ratio:2 attracted:1 bd:1 equilibrium:16 aul:3 sigmoid:2 numerical:1 visible:9 mostafa:6 physical:1 mostly:1 a2:1 negative:1 analog:11 proc:2 design:1 implementation:1 collective:2 tanh:1 amir:1 neuron:5 kosko:2 ith:1 steepest:2 descent:2 dc:1 rather:1 stable:6 lor:1 five:1 afo:1 office:1 consists:2 connection:1 driven:1 store:1 california:1 inequality:2 indeed:1 binary:1 wf:2 arbitrarily:1 trans:1 el:1 lwj:1 pattern:2 bam:1 xm:3 appeared:1 little:1 spurious:1 pasadena:1 hidden:12 increasing:2 jq:1 wij:1 monotonically:1 signal:2 ii:1 full:2 recommended:1 saturation:2 memory:11 force:2 faster:1 eigenvector:1 never:1 having:1 shaped:1 technology:1 ahna:4 identical:1 ae:3 zl:1 unit:7 grant:1 literature:1 randomly:2 acknowledgement:1 before:1 positive:1 engineering:2 limit:2 acad:2 consisting:1 limitation:2 n1:1 cliff:1 sr:1 versus:1 firing:1 might:1 plus:1 englewood:1 multiply:1 db:1 usa:2 basin:2 effectiveness:1 dq:1 vij:2 nl:6 storing:1 smc:1 range:2 enough:1 asymtotically:2 variety:2 supported:1 zi:1 yj:1 fu:1 architecture:3 partial:1 institute:1 empirical:1 distributed:1 feedback:5 dimension:1 xn:1 ul:2 yaser:1 get:1 close:4 aai:1 remark:1 storage:6 prentice:1 cybern:1 far:3 snr:1 demonstrated:1 dz:1 confirm:1 dealing:1 attention:2 atiya:6 stored:6 xi:2 exist:2 percentage:1 zj:1 rule:1 attraction:3 continuous:4 un:6 stability:2 discrete:2 ca:1 ym:1 vol:3 abu:6 target:3 du:3 threshold:3 satisfied:1 element:1 recognition:1 satisfying:1 big:1 derivative:1 ghl:1 asymptotically:3 noise:3 n2:4 syst:1 nothing:1 potential:1 electrical:2 relieve:1 region:2 wj:1 cycle:2 satisfy:2 slow:1 position:1 performed:1 lot:1 mentioned:1 vf:1 ui:1 layer:18 jacobian:2 guaranteed:1 theorem:3 encountered:1 il:1 air:1 accuracy:2 occur:1 variance:1 correspond:2 azn:4 vjj:1 hopfield:6 emergent:1 quantization:2 jersey:1 train:2 nat:2 fast:1 gershgorin:1 department:2 ah:1 submitted:1 zn2:1 vii:1 reach:1 whose:2 valued:2 wi:1 say:1 pp:2 otherwise:1 tested:1 ability:1 e2:1 proof:1 hl:1 inl:1 noisy:1 stop:1 aa:4 associative:9 adjusting:1 taken:1 eigenvalue:5 recall:3 net:1 eventually:1 propose:1 identity:1 lhl:1 fed:1 causing:1 man:1 bidirectional:1 originally:1 dt:3 response:1 away:1 evaluated:2 furthermore:1 az:2 save:2 alternative:1 convergence:3 working:1 requirement:2 existence:1 original:1 in2:1 generating:1 guaranteeing:1 linearize:1 aj:4 aav:1 scientific:1 received:1 giving:1 strong:1
1,163
2,060
The 9 Factor: Relating Distributions on Features to Distributions on Images James M. Coughlan and A. L. Yuille Smith-Kettlewell Eye Research Institute, 2318 Fillmore Street , San Francisco, CA 94115, USA. Tel. (415) 345-2146/2144. Fax. (415) 345-8455. Email: coughlan@ski.org.yuille@ski.org Abstract We describe the g-factor, which relates probability distributions on image features to distributions on the images themselves. The g-factor depends only on our choice of features and lattice quantization and is independent of the training image data. We illustrate the importance of the g-factor by analyzing how the parameters of Markov Random Field (i.e. Gibbs or log-linear) probability models of images are learned from data by maximum likelihood estimation. In particular, we study homogeneous MRF models which learn image distributions in terms of clique potentials corresponding to feature histogram statistics (d. Minimax Entropy Learning (MEL) by Zhu, Wu and Mumford 1997 [11]) . We first use our analysis of the g-factor to determine when the clique potentials decouple for different features . Second, we show that clique potentials can be computed analytically by approximating the g-factor. Third, we demonstrate a connection between this approximation and the Generalized Iterative Scaling algorithm (GIS), due to Darroch and Ratcliff 1972 [2], for calculating potentials. This connection enables us to use GIS to improve our multinomial approximation, using Bethe-Kikuchi[8] approximations to simplify the GIS procedure. We support our analysis by computer simulations. 1 Introduction There has recently been a lot of interest in learning probability models for vision. The most common approach is to learn histograms of filter responses or, equivalently, to learn probability distributions on features (see right panel of figure (1)). See, for example, [6], [5], [4]. (In this paper the features we are considering will be extracted from the image by filters - hence we use the terms "features" and "filters" synonymously. ) An alternative approach, however , is to learn probability distributions on the images themselves. The Minimax Entropy Learning (MEL) theory [11] uses the maximum entropy principle to learn MRF distributions in terms of clique potentials determined by the feature statistics (i.e. histograms of filter responses). (We note that the maximum entropy principle is equivalent to performing maximum likelihood estimation on an MRF whose form is determined by the choice of feature statistics.) When applied to texture modeling it gives a way to unify the filter based approaches (which are often very effective) with the MRF distribution approaches (which are theoretically attractive). ) \ Figure 1: Distributions on images vs. distributions on features. Left and center panels show a natural image and its image gradient magnitude map , respectively. Right panel shows the empirical histogram (i.e. a distribution on a feature) of the image gradient across a dataset of natural images. This feature distribution can be used to create a MRF distribution over images[10]. This paper introduces the g-factor to examine connections between the distribution over images and the distribution over features. As we describe in this paper (see figure (1)), distributions on images and on features can be related by a g-factor (such factors arise in statistical physics, see [3]) . Understanding the g-factor allows us to approximate it in a form that helps explain why the clique potentials learned by MEL take the form that they do as functions of the feature statistics. Moreover, the MEL clique potentials for different features often seem to be decoupled and the g-factor can explain why, and when, this occurs. (I.e. the two clique potentials corresponding to two features A and B are identical whether we learn them jointly or independently). The g-factor is determined only by the form of the features chosen and the spatial lattice and quantization of the image gray-levels. It is completely independent of the training image data. It should be stressed that the choice of image lattice, gray-level quantization and histogram quantization can make a big difference to the g-factor and hence to the probability distributions which are the output of MEL. In Section (2), we briefly review Minimax Entropy Learning. Section (3) introduces the g-factor and determines conditions for when clique potentials are decoupled. In Section (4) we describe a simple approximation which enables us to learn the clique potentials analytically, and in Section (5) we discuss connections between this approximation and the Generalized Iterative Scaling (GIS) algorithm. 2 Minimax Entropy Learning Suppose we have training image data which we assume has been generated by an (unknown) probability distribution PT(X) where x represents an image. Minimax Entropy Learning (MEL) [11] approximates PT(X) by selecting the distribution with maximum entropy constrained by observed feature statistics i(X) = ;fobs. This gives - P(xIA) = >:. ?( ?) e Z [>:] - - ,where A is a parameter chosen such that Lx P(xIA)?>(X) = 'l/Jobs? Or equivalently, so that <910;{[>:] = ;fobs. i We will treat the special case where the statistics are the histogram of a shiftinvariant filter {fi(X) : i = 1, ... , N} , where N is the total number of pixels in the image. So 'l/Ja = ?>a(x) = -tv L~l ba,' i(X) where a = 1, ... , Q indicates the (quantized) Q N filter response values. The potentials become A??>(X) = -tv La=l Li=l A(a)ba,fi(X) = -tv L~l A(fi(X)). Hence P(xl,X) becomes a MRF distribution with clique potentials given by A(fi (x)). This determines a Markov random field with the clique structure given by the filters {fd. ~ ~ MEL also has a feature selection stage based on Minimum Entropy to determine which features to use in the Maximum Entropy Principle. The features are evaluated by computing the entropy - Lx P(xl,X) log P(xl,X) for each choice of features (with small entropies being preferred). A filter pursuit procedure was described to determine which filters/features should be considered (our approximations work for this also). 3 The g-Factor This section defines the g-factor and starts investigating its properties in subsection (3.1). In particular, when, and why, do clique potentials decouple? More precisely, when do the potentials for filters A and B learned simultaneously differ from the potentials for the two filters when they are learned independently? We address these issues by introducing the g-factor g(;f) and the associated distribution Po (;f): (1) x space -----+ iii space GG g(ijiJ = number of images with histogram iii x Figure 2: The g-factor g(;f) counts the number of images x that have statistics ;f. Note that the g-factor depends only on the choice of filters and is independent of the training image data. Here L is the number of grayscale levels of each pixel, so that LN is the total number of possible images. The g-factor is essentially a combinational factor which counts the number of ways that one can obtain statistics ;f, see figure (2). Equivalently, Po is the default distribution on ;f if the images are generated by white noise (i.e. completely random images). We can use the g-factor to compute the induced distribution P(~I'x) on the statistics determined by MEL: A ~~ P(1/1 I'\) = L X 6;;: ~~ 2(-)P(xl'\) 'j','j' x = g( ~)eX.,j; ~ ~, Z[,\] Z[,\] = L ~ X,j; g(1/1)e? . (2) ,j; Observe that both P(~I'x) and log Z[,X] are sufficient for computing the parameters X. The ,X can be found by solving either of the following two (equivalent) equations: ~ ~ ~ ~ 8 10 zrXl ~ L:,j; P(1/1 I,\) 1/1 = 1/1obs, or = 1/1obs, which shows that knowledge of the g-factor A ;X and eX. ,j; are all that is required to do MEL. Observe from equation (2) that we have P(~I'x = 0) = Po(~) . In other words , setting ,X = 0 corresponds to a uniform distribution on the images x. 3.1 Decoupling Filters We now derive an important property of the minimax entropy approach. As mentioned earlier, it often seems that the potentials for filters A and B decouple. In other words, if one applies MEL to two filters A, B simultaneously bv letting ... ....A . . . B...... ....A -B ... "'""'A . . .B . :..t ... 1/1 = (1/1 ,1/1 ), '\ = (,\ ,'\ ), and 1/1obs = (1/1obs ' 1/1obs)' then the solutIOns'\ A , '\ B to the equations: LP(xl,XA , ,XB)(iA(x) , iB(x)) = (~:bs'~!s)' x (3) are the same (approximately) as the solutions to the equations L: x p(xl,XA )iA(x) = ~!s and L: x P(xl,XB)iB(x) = ~!s, see figure (3) for an example. Figure 3: Evidence for decoupling of features. The left and right panels show the clique potentials learned for the features I and I respectively. The solid lines give the potentials when they are learned individually. The dashed lines show the potentials when they are learned simultaneously. Figure courtesy of Prof. Xiuwen Liu, Florida State University. a ax a ay We now show how this decoupling property arises naturally if the g-factor for the two filters factorizes. This factorization, of course, is a property only of the form of the statistics and is completely independent of whether the statistics of the two filters are dependent for the training data. Property I: Suppose we have two sufficient statistics iA(x), iB (x) which are independent on the lattice in the sense that g(~A,~B ) = gA (~A )gB(~B) , then logZ[,XA,,XB] = logZA[,XA] + logZB[,XB] and p(~A,~B ) = pA(~A)pB(~B ). This implies that the parameters XA, XB can be solved from the independent 81ogZ A[XA] _ -A 8 1ogZ B [XB ] _ -B A -A -A -A . - 'ljJobs or L.,j;A P ('ljJ)'ljJ = 'ljJobs' equatwns 8XA - 'ljJobs' 8XB L.,j;B pB(;fB );fB = ;f~s ' A Moreover, the resulting distribution PUC) can be obtained by multiplying the distributions (l/Z A )e XA .,j;A(x) and (l/ZB) eXB.,j;B(x) together. The point here is that the potential terms for the two statistics ;fA,;fB decouple if the phase factor g(;fA,;fB) can be factorized. We conjecture that this is effectively the case for many linear filters used in vision processing. For example, it is plausible that the g- factor for features 0/ and 0/ factorizes - and figure (3) shows that their clique potentials do decouple (approximately). Clearly, if factorization between filters occurs then it gives great simplification to the system. ox 4 oy Approximating the g-factor for a Single Histogram We now consider the case where the statistic is a single histogram. Our aim is to understand why features whose histograms are of stereotypical shape give rise to potentials of the form given by figure (3). Our results , of course, can be directly extended to multiple histograms if the filters decouple, see subsection (3.1). We first describe the approximation and then discuss its relevance for filter pursuit. We rescale the Xvariables by N so that we have: eNX.?(x) eNX .,j; P(X'I-\) = Z[X] , P('ljJ I-\) = g('ljJ) Z[X] , A _ _ (4) We now consider the approximation that the filter responses {Ii} are independent of each other when the images are uniformly distributed. This is the multinomial approximation. (We attempted a related approximation [1] which was less successful.) It implies that we can express the phase factor as being proportional to a multinomial distribution: (nt:) <P = LN N! N1/Jl N1/JQ (N'ljJd!. .. (N'ljJQ)!o ... 0Q ' n (nt:) _ N! N1/Jl N1/JQ (N'ljJd!. .. (N'ljJQ)!Ol "'OQ (5) where L.~= 1 'ljJa = 1 (by definition) and the {o a} are the means of the components Na } with respect to the distribution Po (;f). As we will describe later , the {oa} will be determined by the filters {fi}. See Coughlan and Yuille, in preparation, for details of how to compute the {oa}. 9 TO <p - This approximation enables us to calculate MEL analytically. Theorem With the multinomial approximation the log partition function is: Q log Z[X] = N log L + N log{~= e " a +1og aa } , (6) a=l and the "potentials" P a} can be solved in t erms of the observed data {'ljJobs ,a} to be: \ -- Iog--, 'ljJobs,a a = 1, .. .,Q. Aa (7) Oa Figure 4: Top row: the multinomial approximation. Bottom row: full implementation of MEL (see text). (Left panels) the potentials, (center panels) synthesized images, and (right panels) the difference between the observed histogram (dashed line) and the histogram of the synthesized images (bold line). Filters were d/dx and d/dy. We note that there is an ambiguity Aa r-+ Aa + K where K is an arbitrary number (recall that L~=l 'IjJ(a) = 1). We fix this ambiguity by setting X= 0 if a. = "Jobs. Proof. Direct calculation. Our simulation results show that this simple approximation gives the typical potential forms generated by Markov Chain Monte Carlo (MCMC) algorithms for Minimax Entropy Learning. Compare the multinomial approximation results with those obtained from a full implementation of MEL by the algorithm used in [11], see figure (4). Filter pursuit is required to determine which filters carry most information. MEL [11] prefers filters (statistics) which give rise to low entropy distributions (this is the "Min" part of Minimax). The entropy is given by H(P) = - L xP(xIX) log P(xIX) = log Z[X] - L~=l Aa'IjJa ? For the multinomial approximation this can be computed to be N log L - N L~=l 'ljJa log ~. This gives an intuitive interpretation of feature pursuit: we should prefer filters whose statistical response to the image training data is as large as possible from their responses to uniformly distributed images. This is measured by the Kullback-Leibler divergence L~= l 'ljJa log ~. Recall that if the multinomial approximation is used for multiple filters then we should simply add together the entropies of different filters. 5 Connections to Generalized Iterative Scaling In this section we demonstrate a connection between the multinomial approximation and Generalized Iterative Scaling (GIS)[2]. GIS is an iterative procedure for calculating clique potentials that is guaranteed to converge to the maximum likelihood values of the potentials given the desired empirical filter marginals (e.g. filter histograms). We show that estimating the potentials by the multinomial approximation is equivalent to the estimate obtained after performing the first iteration of GIS. We also outline an efficient procedure that allows us to continue additional GIS iterations to improve upon the multinomial approximation. The GIS procedure calculates a sequence of distributions on the entire image (and is guaranteed to converge to the correct maximum likelihood distribution), with an update rule given by p(t+1)(x) ex P(O)(x)Il~=l{ :F; } <pa(x), where 'lfJit ) =< <Pa(X) >P(t)(x) is the expected histogram for the distribution at time t. This implies that the corresponding clique potential update equation is given by: t +1) = t ) + log 'lfJ~bs - log 'lfJit ). >.i >.i If we initialize GIS so that the initial distribution is the uniform distribution , i.e. p(O) (x) = L -N, then the distribution after one iteration is p(1) (x) ex e2::a <Pa(X) log(1j;~bs /aa) . In other words, the distribution after one iteration is the MEL distribution with clique potential given by the multinomial approximation. (The result can be adapted to the case of multiple filters, as explained in Coughlan and Yuille, in preparation.) We can iterate GIS to improve the estimate of the clique potentials beyond the accuracy of the multinomial approximation. The main difficulty lies in estimating 'lfJit ) for t > 0 (at t = 0 this expectation is just the mean histogram with respect to the uniform distribution, <l:a, which may be calculated efficiently as described in Coughlan and Yuille, in preparation). One way to approximate these expectations is to apply a Bethe-Kikuchi approximation technique [8], used for estimating marginals on Markov Random Fields, to our MEL distribution. Our technique, which was inspired by the Unified Propagation and Scaling Algorithm [7], consists of writing the Bethe free energy [8] for our 2-d image lattice, simplifying it using the shift invariance of the lattice (which enables the algorithm to run swiftly), and using the Convex-Concave Procedure (CCCP) [9] procedure to obtain an iterative update equation to estimate the histogram expectations. The GIS algorithm is then run using these histogram expectations (the results were accurate and did not improve appreciably by using the higher-order Kikuchi free energy approximation). See Coughlan and Yuille, in preparation, for details of this procedure. 6 Discussion This paper describes the g-factor, which depends on the lattice and quantization and is independent of the training image data. Alternatively it can be thought of as being proportional to the distribution of feature responses when the input images are uniformly distributed. We showed that the g-factor can be used to relate probability distributions on features to distributions on images. In particular, we described approximations which, when valid, enable MEL to be computed analytically. In addition, we can determine when the clique potentials for features decouple, and evaluate how informative each feature is. Finally, we establish a connection between the multinomial approximation and GIS, and outline an efficient procedure based on Bethe-Kikuchi approximations that allows us to continue additional GIS iterations to improve upon the multinomial approximation. Acknowledgements We would like to thank Michael Jordan and Yair Weiss for introducing us to Generalized Iterative Scaling and related algorithms. We also thank Anand Rangarajan, Xiuwen Liu, and Song Chun Zhu for helpful conversations. Sabino Ferreira gave useful feedback on the manuscript. This work was supported by the National Institute of Health (NEI) with grant number R01-EY 12691-01. References [1] J.M. Coughlan and A.L. Yuille. "A Phase Space Approach to Minimax Entropy Learning and The Minutemax approximation". In Proceedings NIPS '98. 1998. [2] J. N. Darroch and D. Ratcliff. "Generalized Iterative Scaling for Log-Linear Models". The Annals of Mathematical Statistics. 1972. Vol. 43, No.5, 14701480. [3] C. Domb and M.S. Green (Eds). Phase Transitions and Critical Phenomena. Vol. 2. Academic Press. London. 1972. [4] S. M. Konishi, A.L. Yuille, J.M. Coughlan and Song Chun Zhu. "Fundamental Bounds on Edge Detection: An Information Theoretic Evaluation of Different Edge Cues." In Proceedings Computer Vision and Pattern Recognition CVPR'99. Fort Collins, Colorado. June 1999. [5] A.B. Lee, D.B. Mumford, and J. Huang. "Occlusion Models of Natural Images: A Statistical Study of a Scale-Invariant Dead Leaf Model". International Journal of Computer Vision. Vol. 41, No.'s 1/2. January/February 2001. [6] J. Portilla and E. P. Simoncelli. "Parametric Texture Model based on Joint Statistics of Complex Wavelet Coefficients" . International Journal of Computer Vision. October 2000. [7] Y. W. Teh and M. Welling. "The Unified Propagation and Scaling Algorithm." In Proceedings NIPS'01. 2001. [8] J.S. Yedidia, W.T. Freeman, Y. Weiss, "Generalized Belief Propagation." In Proceedings NIPS'OO. 2000. [9] A.L. Yuille. "CCCP Algorithms to Minimize the Bethe and Kikuchi Free Energies," Neural Computation. In press. 2002. [10] S.C. Zhu and D. Mumford. "Prior Learning and Gibbs Reaction-Diffusion." PAMI vo1.19, no.11, pp1236-1250, Nov. 1997. [11] S.C. Zhu, Y. Wu, and D. Mumford. "Minimax Entropy Principle and Its Application to Texture Modeling". Neural Computation. Vol. 9. no. 8. Nov. 1997.
2060 |@word briefly:1 seems:1 simulation:2 simplifying:1 solid:1 carry:1 initial:1 liu:2 selecting:1 reaction:1 nt:2 erms:1 dx:1 partition:1 informative:1 shape:1 enables:4 update:3 v:1 cue:1 leaf:1 coughlan:8 smith:1 quantized:1 lx:2 org:2 mathematical:1 direct:1 become:1 kettlewell:1 consists:1 combinational:1 theoretically:1 expected:1 themselves:2 examine:1 ol:1 inspired:1 freeman:1 considering:1 becomes:1 estimating:3 moreover:2 panel:7 factorized:1 unified:2 concave:1 ferreira:1 grant:1 treat:1 analyzing:1 approximately:2 pami:1 factorization:2 procedure:9 logz:1 empirical:2 thought:1 word:3 ga:1 selection:1 writing:1 equivalent:3 map:1 center:2 courtesy:1 independently:2 convex:1 unify:1 rule:1 stereotypical:1 konishi:1 annals:1 pt:2 suppose:2 colorado:1 enx:2 homogeneous:1 us:1 pa:4 recognition:1 observed:3 bottom:1 solved:2 calculate:1 mentioned:1 solving:1 yuille:9 upon:2 completely:3 po:4 joint:1 describe:5 effective:1 monte:1 london:1 whose:3 plausible:1 cvpr:1 statistic:17 gi:14 jointly:1 sequence:1 intuitive:1 rangarajan:1 kikuchi:5 help:1 illustrate:1 derive:1 oo:1 measured:1 rescale:1 job:2 implies:3 differ:1 correct:1 lfj:1 filter:34 enable:1 ja:1 fix:1 considered:1 great:1 estimation:2 individually:1 appreciably:1 create:1 clearly:1 aim:1 factorizes:2 og:1 ax:1 june:1 likelihood:4 ratcliff:2 indicates:1 sense:1 helpful:1 dependent:1 entire:1 jq:2 pixel:2 issue:1 spatial:1 special:1 constrained:1 initialize:1 field:3 ijij:1 identical:1 represents:1 synonymously:1 simplify:1 simultaneously:3 divergence:1 national:1 phase:4 occlusion:1 n1:4 detection:1 interest:1 fd:1 evaluation:1 introduces:2 xb:7 chain:1 accurate:1 edge:2 decoupled:2 desired:1 modeling:2 earlier:1 lattice:7 introducing:2 uniform:3 successful:1 minutemax:1 swiftly:1 fundamental:1 international:2 lee:1 physic:1 michael:1 together:2 na:1 ambiguity:2 huang:1 dead:1 li:1 potential:32 domb:1 bold:1 coefficient:1 depends:3 later:1 lot:1 start:1 minimize:1 il:1 accuracy:1 efficiently:1 carlo:1 multiplying:1 explain:2 ed:1 email:1 definition:1 energy:3 james:1 e2:1 naturally:1 associated:1 proof:1 dataset:1 recall:2 subsection:2 knowledge:1 conversation:1 manuscript:1 higher:1 response:7 wei:2 evaluated:1 ox:1 xa:8 stage:1 just:1 propagation:3 defines:1 gray:2 usa:1 analytically:4 hence:3 leibler:1 white:1 attractive:1 mel:17 generalized:7 gg:1 ay:1 outline:2 demonstrate:2 theoretic:1 image:40 recently:1 fi:5 common:1 multinomial:15 jl:2 interpretation:1 approximates:1 relating:1 synthesized:2 marginals:2 gibbs:2 add:1 showed:1 continue:2 minimum:1 additional:2 ey:1 determine:5 converge:2 dashed:2 ii:1 relates:1 multiple:3 full:2 simoncelli:1 academic:1 calculation:1 cccp:2 iog:1 calculates:1 mrf:6 vision:5 essentially:1 expectation:4 histogram:18 iteration:5 addition:1 induced:1 anand:1 oq:1 seem:1 jordan:1 iii:2 iterate:1 gave:1 shift:1 whether:2 darroch:2 gb:1 fob:2 song:2 prefers:1 useful:1 vol:4 express:1 pb:2 diffusion:1 run:2 wu:2 ob:5 dy:1 scaling:8 prefer:1 bound:1 guaranteed:2 simplification:1 bv:1 adapted:1 precisely:1 min:1 performing:2 conjecture:1 tv:3 across:1 describes:1 lp:1 b:3 explained:1 invariant:1 ln:2 equation:6 discus:2 count:2 letting:1 pursuit:4 yedidia:1 apply:1 observe:2 alternative:1 yair:1 florida:1 top:1 calculating:2 prof:1 establish:1 approximating:2 february:1 r01:1 occurs:2 mumford:4 fa:2 parametric:1 gradient:2 thank:2 oa:3 street:1 equivalently:3 october:1 relate:1 rise:2 ba:2 implementation:2 ski:2 unknown:1 teh:1 xiuwen:2 markov:4 january:1 extended:1 portilla:1 arbitrary:1 nei:1 fort:1 required:2 connection:7 learned:7 nip:3 address:1 beyond:1 pattern:1 green:1 belief:1 ia:3 critical:1 natural:3 difficulty:1 zhu:5 minimax:10 improve:5 eye:1 health:1 fax:1 ljj:4 text:1 review:1 understanding:1 acknowledgement:1 prior:1 oy:1 proportional:2 sufficient:2 xp:1 principle:4 row:2 course:2 supported:1 free:3 understand:1 institute:2 distributed:3 xix:2 xia:2 default:1 calculated:1 valid:1 feedback:1 transition:1 fb:4 san:1 welling:1 approximate:2 nov:2 preferred:1 kullback:1 clique:19 investigating:1 francisco:1 alternatively:1 grayscale:1 iterative:8 why:4 learn:7 bethe:5 ca:1 decoupling:3 tel:1 complex:1 did:1 main:1 big:1 noise:1 arise:1 fillmore:1 xl:7 lie:1 ib:3 third:1 wavelet:1 theorem:1 chun:2 evidence:1 quantization:5 effectively:1 importance:1 texture:3 magnitude:1 entropy:19 simply:1 ijj:1 applies:1 aa:6 corresponds:1 determines:2 extracted:1 determined:5 typical:1 uniformly:3 vo1:1 decouple:7 zb:1 total:2 invariance:1 la:1 attempted:1 shiftinvariant:1 support:1 stressed:1 arises:1 collins:1 relevance:1 preparation:4 evaluate:1 mcmc:1 phenomenon:1 ex:4
1,164
2,061
Small-World Phenomena and the Dynamics of Information Jon Kleinberg Department of Computer Science Cornell University Ithaca NY 14853 1 Introduction The problem of searching for information in networks like the World Wide Web can be approached in a variety of ways, ranging from centralized indexing schemes to decentralized mechanisms that navigate the underlying network without knowledge of its global structure. The decentralized approach appears in a variety of settings: in the behavior of users browsing the Web by following hyperlinks; in the design of focused crawlers [4, 5, 8] and other agents that explore the Web?s links to gather information; and in the search protocols underlying decentralized peer-to-peer systems such as Gnutella [10], Freenet [7], and recent research prototypes [21, 22, 23], through which users can share resources without a central server. In recent work, we have been investigating the problem of decentralized search in large information networks [14, 15]. Our initial motivation was an experiment that dealt directly with the search problem in a decidedly pre-Internet context: Stanley Milgram?s famous study of the small-world phenomenon [16, 17]. Milgram was seeking to determine whether most pairs of people in society were linked by short chains of acquaintances, and for this purpose he recruited individuals to try forwarding a letter to a designated ?target? through people they knew on a firstname basis. The starting individuals were given basic information about the target ? his name, address, occupation, and a few other personal details ? and had to choose a single acquaintance to send the letter to, with goal of reaching the target as quickly as possible; subsequent recipients followed the same procedure, and the chain closed in on its destination. Of the chains that completed, the median number of steps required was six ? a result that has since entered popular culture as the ?six degrees of separation? principle [11]. Milgram?s experiment contains two striking discoveries ? that short chains are pervasive, and that people are able to find them. This latter point is concerned precisely with a type of decentralized navigation in a social network, consisting of people as nodes and links joining pairs of people who know each other. From an algorithmic perspective, it is an interesting question to understand the structure of networks in which this phenomenon emerges ? in which message-passing with purely local information can be efficient. Networks that Support Efficient Search. A model of a ?navigable? network requires a few basic features. It should contain short paths among all (or most) pairs of nodes. To be non-trivial, its structure should be partially known and partially unknown to its constituent nodes; in this way, information about the known parts can be used to construct paths that make use of the unknown parts as well. This is clearly what was taking place in Milgram?s experiments: participants, using the information available to them, were estimating which of their acquaintances would lead to the shortest path through the full social network. Guided by these observations, we turned to the work of Watts and Strogatz [25], which proposes a model of ?small-world networks? that very concisely incorporates these features. A simple variant of their basic model can be described as follows. One starts with a p-dimensional lattice, in which nodes are joined only to their nearest neighbors. One then adds k directed long-range links out of each node v, for a constant k; the endpoint of each link is chosen uniformly at random. Results from the theory of random graphs can be used to show that with high probability, there will be short paths connecting all pairs of nodes (see e.g. [3]); at the same time, the network will locally retain a lattice-like structure. Asymptotically, our criterion for ?shortness? of paths is what one obtains from this and similar random constructions: there should be paths whose lengths are bounded by a polynomial in log n, where n is the number of nodes. We will refer to such a function as polylogarithmic. This network model, a superposition of a lattice and a set of long-range links, is a natural one in which to study the behavior of a decentralized search algorithm. The algorithm knows the structure of the lattice; it starts from a node s, and is told the coordinates of a target node t. It successively traverses links of the network so as to reach the target as quickly as possible; but, crucially, it does not know the long-range links out of any node that it has not yet visited. In addition to moving forward across directed links, the algorithm may travel in reverse across any link that it has already followed in the forward direction; this allows it to back up when it does not want to continue exploring from its current node. One can view this as hitting the ?back button? on a Web browser ? or returning the letter to its previous holder in Milgram?s experiments, with instructions that he or she should try someone else. We say that the algorithm has search time Y (n) if, on a randomly generated n-node network with s and t chosen uniformly at random, it reaches the target t in at most Y (n) steps with probability at least 1 ? ?(n), for a function ?(?) going to 0 with n. The first result in [15] is that no decentralized algorithm can achieve a polylogarithmic search time in this network model ? even though, with high probability, there are paths of polylogarithmic length joining all pairs of nodes. However, if we generalize the model slightly, then it can support efficient search. Specifically, when we construct a long-range link (v, w) out of v, we do not choose w uniformly at random; rather, we choose it with probability proportional to d??, where d is the lattice distance from v to w. In this way, the long-range links become correlated to the geometry of the lattice. We show in [15] that when ? is equal to p, the dimension of the underlying lattice, then a decentralized greedy algorithm achieves search time proportional to log2 n; and for any other value of ?, there is no decentralized algorithm with a polylogarithmic search time. Recent work by Zhang, Goel, and Govindan [26] has shown how the distribution of links associated with the optimal value of ? may lead to improved performance for decentralized search in the Freenet peer-to-peer system. Adamic, Lukose, Puniyani, and Huberman [2] have recently considered a variation of the decentralized search problem in a network that has essentially no known underlying structure; however, when the number of links incident to nodes follows a power-law distribution, then a search strategy that seeks high-degree nodes can be effective. They have applied their results to the Gnutella system, which exhibits such a structure. In joint work with Kempe and Demers [12], we have studied how distributions that are inverse-polynomial in the distance between nodes can be used in the design of gossip protocols for spreading information in a network of communicating agents. The goal of the present paper is to consider more generally the problem of decentralized search in networks with partial information about the underlying structure. While a lattice makes for a natural network backbone, we would like to understand the extent to which the principles underlying efficient decentralized search can be abstracted away from a lattice-like structure. We begin by considering networks that are generated from a hierarchical structure, and show that qualitatively similar results can be obtained; we then discuss a general model of group structures, which can be viewed as a simultaneous generalization of lattices and hierarchies. We refer to k, the number of out-links per node, as the out-degree of the model. The technical details of our results ? both in the statements of the results and the proofs ? are simpler when we allow the out-degree to be polylogarithmic, rather than constant. Thus we describe this case first, and then move on to the case in which each node has only a constant number of out-links. 2 Hierarchical Network Models In a number of settings, nodes represent objects that can be classified according to a hierarchy or taxonomy; and nodes are more likely to form links if they belong to the same small sub-tree in the hierarchy, indicating they are more closely related. To construct a network model from this idea, we represent the hierarchy using a complete b-ary tree T , where b is a constant. Let V denote the set of leaves of T ; we use n to denote the size of V , and for two leaves v and w, we use h(v, w) to denote the height of the least common ancestor of v and w in T . We are also given a monotone non-increasing function f(?) that will determine link probabilities. For each node v ? V , we create a random link to w with probability proportional P to f(h(v, w)); in other words, the probability of choosing w is equal to f(h(v, w))/ x6=v f(h(v, x)). We create k links out of each node v this way, choosing the endpoint w each time independently and with repetition allowed. This results in a graph G on the set V . For the analysis in this section, we will take the out-degree to be k = c log2 n, for a constant c. It is important to note that the tree T is used only in the generation process of G; neither the edges nor the non-leaf nodes of T appear in G. (By way of contrast, the lattice model in [15] included both the long-range links and the nearest-neighbor links of the lattice itself.) When we use the term ?node? without any qualification, we are referring to nodes of G, and hence to leaves of T ; we will use ?internal node? to refer to non-leaf nodes of T . We refer to the process producing G as a hierarchical model with exponent ? if the function f(h) grows asymptotically like b??h: 00 f(h) b?? h 0 = 0 for all ?00 > ?. 0 h = 0 for all ? < ? and lim ?? h?? b h?? f(h) lim There are several natural interpretations for a hierarchical network model. One is in terms of the World Wide Web, where T is a topic hierarchy such as www.yahoo.com. Each leaf of T corresponds to a Web page, and its path from the root specifies an increasingly fine-grained description of the page?s topic. Thus, a particular leaf may be associated with Science/Computer Science/Algorithms or with Arts/Music/Opera. The linkage probabilities then have a simple meaning ? they are based on the distance between the topics of the pages, as measured by the height of their least common ancestor in the topic hierarchy. A page on Sci- ence/Computer Science/Algorithms may thus be more likely to link to one on Science/Computer Science/Machine Learning than to one on Arts/Music/Opera. Of course, the model is a strong simplification, since topic structures are not fully hierarchical, and certainly do not have uniform branching and depth. It is worth noting that a number of recent models for the link structure of the Web, as well as other relational structures, have looked at different ways in which similarities in content can affect the probability of linkage; see e.g. [1, 6, 9]. Another interpretation of the hierarchical model is in terms of Milgram?s original experiment. Studies performed by Killworth and Bernard [13] showed that in choosing a recipient for the letter, participants were overwhelmingly guided by one of two criteria: similarity to the target in geography, or similarity to the target in occupation. If one views the lattice as forming a simple model for geographic factors, the hierarchical model can similarly be interpreted as forming a ?topic hierarchy? of occupations, with individuals at the leaves. Thus, for example, the occupations of ?banker? and ?stock broker? may belong to the same small sub-tree; since the target in one of Milgram?s experiments was a stock broker, it might therefore be a good strategy to send the letter to a banker. Independently of our work here, Watts, Dodds, and Newman have recently studied hierarchical structures for modeling Milgram?s experiment in social networks [24]. We now consider the search problem in a graph G generated from a hierarchical model: A decentralized algorithm has knowledge of the tree T , and knows the location of a target leaf that it must reach; however, it only learns the structure of G as it visits nodes. The exponent ? determines how the structures of G and T are related; how does this affect the navigability of G? In the analysis of the lattice model [15], the key property of the optimal exponent was that, from any point, there was a reasonable probability of a long-range link that halved the distance to the target. We make use of a similar idea here: when ? = 1, there is always a reasonable probability of finding a long-range link into a strictly smaller sub-tree containing the target. As mentioned above, we focus here on the case of polylogarithmic outdegree, with the case of constant out-degree deferred until later. Theorem 2.1 (a) There is a hierarchical model with exponent ? = 1 and polylogarithmic out-degree in which a decentralized algorithm can achieve search time O(log n). (b) For every ? 6= 1, there is no hierarchical model with exponent ? and polylogarithmic out-degree in which a decentralized algorithm can achieve polylogarithmic search time. Due to space limitations, we omit proofs from this version of the paper. Complete proofs may be found in the extended version, which is available on the author?s Web page (http://www.cs.cornell.edu/home/kleinber/). To prove (a), we show that when the search is at a node v whose least common ancestor with the target has height h, there is a high probability that v has a link into the sub-tree of height h?1 containing the target. In this way, the search reaches the target in logarithmically many steps. To prove (b), we exhibit a sub-tree T 0 containing the target such that, with high probability, it takes any decentralized algorithm more than a polylogarithmic number of steps to find a link into T 0 . 3 Group Structures The analysis of the search problem in a hierarchical model is similar to the analysis of the lattice-based approach in [15], although the two types of models seem superficially quite different. It is natural to look for a model that would serve as a simultaneous generalization of each. Consider a collection of individuals in a social network, and suppose that we know of certain groups to which individuals belong ? people who live in the same town, or work in the same profession, or have some other affiliation in common. We could imagine that people are more likely to be connected if they both belong to the same small group. In a lattice-based model, there may be a group for each subset of lattice points contained in a common ball (grouping based on proximity); in a hierarchical model, there may be a group for each subset of leaves contained in a common sub-tree. We now discuss the notion of a group structure, to make this precise; we follow a model proposed in joint work with Kempe and Demers [12], where we were concerned with designing gossip protocols for lattices and hierarchies. A technically different model of affiliation networks, also motivated by these types of issues, has been studied recently by Newman, Watts, and Strogatz [18]. A group structure consists of an underlying set V of nodes, and a collection of subsets of V (the groups). The collection of groups must include V itself; and it must satisfy the following two properties, for constants ? < 1 and ? > 1. (i) If R is a group of size q ? 2 containing a node v, then there is a group R0 ? R containing v that is strictly smaller than R, but has size at least ?q. (ii) If R1, R2, R3, . . . are groups that all have size at most q and all contain a common node v, then their union has size at most ?q. The reader can verify that these two properties hold for the collection of balls in a lattice, as well as for the collection of sub-trees in a hierarchy. However, it is easy to construct examples of group structures that do not arise in this way from lattices or hierarchies. Given a group structure (V, {Ri}), and a monotone non-increasing function f(?), we now consider the following process for generating a graph on V . For two nodes v and w, we use q(v, w) to denote the minimum size of a group containing both v and w. (Note that such a group must exist, since V itself is a group.) For each node v ? V , we create a random link to w with probability proportional to f(q(v, w)); repeating this k times independently yields k links out of v. We refer to this as a group-induced model with exponent ? if f(q) grows asymptotically like q ?? : 00 lim h?? f(q) q?? 0 = 0 for all ? < ? and lim = 0 for all ?00 > ?. h?? f(q) q??0 A decentralized search algorithm in such a network is given knowledge of the full group structure, and must follow links of G to a designated target t. We now state an analogue of Theorem 2.1 for group structures. Theorem 3.1 (a) For every group structure, there is a group-induced model with exponent ? = 1 and polylogarithmic out-degree in which a decentralized algorithm can achieve search time O(log n). (b) For every ? < 1, there is no group-induced model with exponent ? and polylogarithmic out-degree in which a decentralized algorithm can achieve polylogarithmic search time. Notice that in a hierarchical model, the smallest group (sub-tree) containing two nodes v and w has size bh(v,w) , and so Theorem 3.1(a) implies Theorem 2.1(a). Similarly, on a lattice, the smallest group (ball) containing two nodes v and w at lattice distance d has size ?(dp ), and so Theorem 3.1(a) implies a version of the result from [15], that efficient search is possible in a lattice model when nodes form links with probability d?p . (In the version of the lattice result implied here, there are no nearest-neighbor links at all; but each node has a polylogarithmic number of out-links.) The proof of Theorem 3.1(a) closely follows the proof of Theorem 2.1(a). We consider a node v ? the current point in the search ? for which the smallest group containing v and the target t has size q. Using group structure properties (i) and (ii), we show there is a high probability that v has a link into a group containing t of size between ?2 q and ?q. In this way, the search passes through groups containing t of sizes that diminish geometrically, and hence it terminates in logarithmic time. Note that Theorem 3.1(b) only considers exponents ? < 1. This is because there exist group-induced models with exponents ? > 1 in which decentralized algorithms can achieve polylogarithmic search time. For example, consider an undirected graph G? in which each node has 3 neighbors, and each pair of nodes can be connected by a path of length O(log n). It is possible to define a group structure satisfying properties (i) and (ii) in which each edge of G? appears as a 2-node group; but then, a graph G generated from a group-induced model with a very large exponent ? will contain all edges of G? with high probability, and a decentralized search algorithm will be able to follow these edges directly to construct a short path to the target. However, a lower bound for the case ? > 1 can be obtained if we place one additional restriction on the group structure. Give a group structure (V, {Ri}), and a cut-off value q, we define a graph H(q) on V by joining any two nodes that belong to a common group of size at most q. Note that H(q) is not a random graph; it is defined simply in terms of the group structure and q. We now argue that if many pairs of nodes are far apart in H(q), for a suitably large value of q, then a decentralized algorithm cannot be efficient when ? > 1. Theorem 3.2 Let (V, {Ri }) be a group structure. Suppose there exist constants ?, ? > 0 so that a constant fraction of all pairs of nodes have shortest-path distance ?(n? ) in H(n? ). Then for every ? > 1, there is no group-induced model on (V, {R i}) with exponent ? and a polylogarithmic number of out-links per node in which a decentralized algorithm can achieve polylogarithmic search time. Notice this property holds for group structures arising from both lattices and hierarchies; in a lattice, a constant fraction of all pairs in H(n1/2p) have distance ?(n1/2p), while in a hierarchy, the graph H(n? ) is disconnected for every ? < 1. 4 Nodes with a Constant Number of Out-Links Thus far, by giving each node more than a constant number of out-links, we have been able to design very simple search algorithms in networks generated according to the optimal exponent ?. From each node, there is a way to make progress toward the target node t, and so the structure of the graph G funnels the search towards its destination. When the out-degree is constant, however, things get much more complicated. First of all, with high probability, many nodes will have all their links leading ?away? from the target in the hierarchy. Second, there is a constant probability that the target t will have no in-coming links, and so the whole task of finding t becomes ill-defined. This indicates that the statement of the results themselves in this case will have to be somewhat different. In this section, we work with a hierarchical model, and construct graphs with con- stant out-degree k; the value of k will need to be sufficiently large in terms of other parameters of the model. It is straightforward to formulate an analogue of our results for group structures, but we do not go into the details of this here. To deal with the problem that t itself may have no incoming links, we relax the search problem to that of finding a cluster of nodes containing t. In a topic-based model of Web pages, for example, we can consider t as a representative of a desired type of page, with goal being to find any page of this type. Thus, we are given a complete b-ary tree T , where b is a constant; we let L denote the set of leaves of T , and m denote the size of L. We place r nodes at each leaf of T , forming a set V of n = mr nodes total. We then define a graph G on V as in Section 2: for a non-increasing function f(?), we create k links out of each node v ? V , choosing w as an endpoint with probability proportional to f(h(v, w)). As before, we refer to this process as a hierarchical model with exponent ?, for the appropriate value of ?. We refer to each set of r nodes at a common leaf of T as a cluster, and define the resolution of the hierarchical model to be the value r. A decentralized algorithm is given knowledge of T , and a target node t; it must reach any node in the cluster containing t. Unlike the previous algorithms we have developed, which only moved forward across links, the algorithm we design here will need to make use of the ability to travel in reverse across any link that it has already followed in the forward direction. Note also that we cannot easily reduce the current search problem to that of Section 2 by collapsing clusters into ?super-nodes,? since there are not necessarily links joining nodes within the same cluster. The search task clearly becomes easier as the resolution of the model (i.e. the size of clusters) becomes larger. Thus, our goal is to achieve polylogarithmic search time in a hierarchical model with polylogarithmic resolution. Theorem 4.1 (a) There is a hierarchical model with exponent ? = 1, constant out-degree, and polylogarithmic resolution in which a decentralized algorithm can achieve polylogarithmic search time. (b) For every ? 6= 1, there is no hierarchical model with exponent ?, constant outdegree, and polylogarithmic resolution in which a decentralized algorithm can achieve polylogarithmic search time. The search algorithm used to establish part (a) operates in phases. It begins each phase j with a collection of ?(log n) nodes all belonging to the sub-tree Tj that contains the target t and whose root is at depth j. During phase j, it explores outward from each of these nodes until it has discovered a larger but still polylogarithmicsized set of nodes belonging to Tj . From among these, there is a high probability that at least ?(log n) have links into the smaller sub-tree Tj+1 that contains t and whose root is at depth j + 1. At this point, phase j + 1 begins, and the process continues until the cluster containing t is found. Acknowledgments My thinking about models for Web graphs and social networks has benefited greatly from discussions and collaboration with Dimitris Achlioptas, Avrim Blum, Duncan Callaway, Michelle Girvan, John Hopcroft, David Kempe, Ravi Kumar, Tom Leighton, Mark Newman, Prabhakar Raghavan, Sridhar Rajagopalan, Steve Strogatz, Andrew Tomkins, Eli Upfal, and Duncan Watts. The research described here was supported in part by a David and Lucile Packard Foundation Fellowship, an ONR Young Investigator Award, NSF ITR/IM Grant IIS-0081334, and NSF Faculty Early Career Development Award CCR-9701399. References [1] D. Achlioptas, A. Fiat, A. Karlin, F. McSherry, ?Web search via hub synthesis,? Proc. 42nd IEEE Symp. on Foundations of Computer Science, 2001. [2] L. Adamic, R. Lukose, A. Puniyani, B. Huberman, ?Search in Power-Law Networks,? Phys. Rev. E, 64 46135 (2001) [3] B. Bollob? as, F. Chung, ?The diameter of a cycle plus a random matching,? SIAM J. Disc. Math. 1(1988). [4] S. Chakrabarti, M. van den Berg, B. Dom, ?Focused crawling: A new approach to topic-specific Web resource discovery,? Proc. 8th Intl. World Wide Web Conf., 1999. [5] J. Cho, H. Garcia-Molina, L. Page, ?Efficient Crawling Through URL Ordering,? Proc. 7th Intl. World Wide Web Conf., 1998. [6] D. Cohn and T. Hofmann, ?The Missing Link ? A Probabilistic Model of Document Content and Hypertext Connectivity,? Adv. Neural Inf. Proc. Sys. (NIPS) 13, 2000. [7] I. Clarke, O. Sandberg, B. Wiley, T. Hong, ?Freenet: A Distributed Anonymous Information Storage and Retrieval System,? International Workshop on Design Issues in Anonymity and Unobservability, 2000. [8] M. Diligenti, F.M. Coetzee, S. Lawrence, C.L. Giles, M. Gori, ?Focused Crawling Using Context Graphs,? Proc. 26th Intl. Conf. on Very Large Databases (VLDB), 2000. [9] L. Getoor, N. Friedman, D. Koller, and B. Taskar. ?Learning Probabilistic Models of Relational Structure,? Proc. 18th International Conference on Machine Learning, 2001. [10] Gnutella. http://gnutella.wego.com. [11] J. Guare, Six Degrees of Separation: A Play (Vintage Books, New York, 1990). [12] D. Kempe, J. Kleinberg, A. Demers. ?Spatial gossip and resource location protocols,? Proc. 33rd ACM Symp. on Theory of Computing, 2001. [13] P. Killworth, H. Bernard, ?Reverse small world experiment,? Social Networks 1(1978). [14] J. Kleinberg. ?Navigation in a Small World.? Nature 406(2000). [15] J. Kleinberg. ?The small-world phenomenon: An algorithmic perspective.? Proc. 32nd ACM Symposium on Theory of Computing, 2000. Also appears as Cornell Computer Science Technical Report 99-1776 (October 1999). [16] M. Kochen, Ed., The Small World (Ablex, Norwood, 1989). [17] S. Milgram, ?The small world problem,? Psychology Today 1(1967). [18] M. Newman, D. Watts, S. Strogatz, ?Random graph models of social networks,? Proc. Natl. Acad. Sci., to appear. [19] A. Oram, editor, Peer-to-Peer: Harnessing the Power of Disruptive Technologies O?Reilly and Associates, 2001. [20] A. Puniyani, R. Lukose, B. Huberman, ?Intentional Walks on Scale Free Small Worlds,? HP Labs Information Dynamics Group, at http://www.hpl.hp.com/shl/. [21] S. Ratnasamy, P. Francis, M. Handley, R. Karp, S. Shenker, ?A Scalable ContentAddressable Network,? Proc. ACM SIGCOMM, 2001 [22] A. Rowstron, P. Druschel, ?Pastry: Scalable, distributed object location and routing for large-scale peer-to-peer systems,? Proc. 18th IFIP/ACM International Conference on Distributed Systems Platforms (Middleware 2001), 2001. [23] I. Stoica, R. Morris, D. Karger, F. Kaashoek, H. Balakrishnan, ?Chord: A Scalable Peer-to-peer Lookup Service for Internet Applications,? Proc. ACM SIGCOMM, 2001 [24] D. Watts, P. Dodds, M. Newman, personal communication, December 2001. [25] D. Watts, S. Strogatz, ?Collective dynamics of small-world networks,? Nature 393(1998). [26] H. Zhang, A. Goel, R. Govindan, ?Using the Small-World Model to Improve Freenet Performance,? Proc. IEEE Infocom, 2002.
2061 |@word faculty:1 version:4 polynomial:2 leighton:1 nd:2 suitably:1 instruction:1 vldb:1 seek:1 crucially:1 initial:1 contains:3 karger:1 document:1 current:3 com:3 yet:1 crawling:3 must:6 john:1 subsequent:1 hofmann:1 greedy:1 leaf:13 sys:1 short:5 math:1 node:63 location:3 traverse:1 simpler:1 zhang:2 height:4 become:1 symposium:1 chakrabarti:1 prove:2 consists:1 symp:2 behavior:2 themselves:1 nor:1 considering:1 increasing:3 becomes:3 begin:3 estimating:1 underlying:7 bounded:1 what:2 backbone:1 interpreted:1 navigability:1 developed:1 finding:3 every:6 returning:1 grant:1 omit:1 appear:2 producing:1 before:1 service:1 local:1 qualification:1 acad:1 joining:4 path:11 might:1 plus:1 studied:3 forwarding:1 someone:1 callaway:1 range:8 directed:2 acknowledgment:1 union:1 procedure:1 matching:1 reilly:1 pre:1 word:1 get:1 cannot:2 bh:1 storage:1 context:2 live:1 www:3 restriction:1 missing:1 send:2 straightforward:1 go:1 starting:1 independently:3 focused:3 formulate:1 resolution:5 communicating:1 his:1 searching:1 notion:1 coordinate:1 variation:1 target:24 construction:1 hierarchy:13 user:2 suppose:2 imagine:1 play:1 today:1 designing:1 associate:1 logarithmically:1 satisfying:1 anonymity:1 continues:1 cut:1 database:1 taskar:1 hypertext:1 connected:2 cycle:1 adv:1 ordering:1 chord:1 mentioned:1 rowstron:1 dynamic:3 personal:2 dom:1 ablex:1 purely:1 serve:1 technically:1 basis:1 easily:1 joint:2 hopcroft:1 stock:2 druschel:1 effective:1 describe:1 approached:1 newman:5 choosing:4 harnessing:1 peer:10 whose:4 quite:1 larger:2 say:1 relax:1 ability:1 browser:1 itself:4 karlin:1 coming:1 turned:1 entered:1 achieve:10 description:1 moved:1 constituent:1 cluster:7 r1:1 prabhakar:1 intl:3 generating:1 object:2 andrew:1 measured:1 nearest:3 progress:1 strong:1 c:1 implies:2 direction:2 guided:2 closely:2 raghavan:1 routing:1 generalization:2 geography:1 anonymous:1 im:1 exploring:1 strictly:2 hold:2 proximity:1 sufficiently:1 considered:1 diminish:1 intentional:1 lawrence:1 algorithmic:2 achieves:1 early:1 smallest:3 purpose:1 proc:13 travel:2 spreading:1 superposition:1 visited:1 repetition:1 create:4 pastry:1 clearly:2 always:1 super:1 reaching:1 rather:2 cornell:3 karp:1 overwhelmingly:1 pervasive:1 focus:1 she:1 indicates:1 greatly:1 contrast:1 koller:1 ancestor:3 going:1 issue:2 among:2 ill:1 exponent:16 yahoo:1 proposes:1 development:1 art:2 spatial:1 kempe:4 platform:1 equal:2 construct:6 shl:1 outdegree:2 dodds:2 look:1 jon:1 thinking:1 report:1 few:2 randomly:1 individual:5 geometry:1 consisting:1 phase:4 n1:2 friedman:1 centralized:1 message:1 lucile:1 certainly:1 deferred:1 navigation:2 tj:3 mcsherry:1 natl:1 chain:4 edge:4 partial:1 culture:1 tree:14 walk:1 desired:1 modeling:1 giles:1 ence:1 lattice:26 subset:3 uniform:1 my:1 cho:1 referring:1 explores:1 siam:1 international:3 ifip:1 retain:1 destination:2 told:1 off:1 probabilistic:2 connecting:1 synthesis:1 quickly:2 connectivity:1 central:1 town:1 successively:1 choose:3 containing:14 collapsing:1 conf:3 book:1 acquaintance:3 chung:1 leading:1 lookup:1 satisfy:1 stoica:1 performed:1 try:2 closed:1 view:2 linked:1 root:3 later:1 start:2 francis:1 participant:2 complicated:1 lab:1 opera:2 holder:1 who:2 yield:1 dealt:1 generalize:1 famous:1 disc:1 worth:1 ary:2 classified:1 simultaneous:2 reach:5 phys:1 milgram:9 ed:1 associated:2 proof:5 coetzee:1 con:1 oram:1 navigable:1 demers:3 popular:1 knowledge:4 emerges:1 lim:4 stanley:1 fiat:1 vintage:1 profession:1 sandberg:1 back:2 appears:3 steve:1 x6:1 follow:3 tom:1 improved:1 though:1 achlioptas:2 until:3 hpl:1 web:14 adamic:2 cohn:1 grows:2 name:1 contain:3 geographic:1 verify:1 hence:2 deal:1 during:1 branching:1 criterion:2 hong:1 complete:3 ranging:1 meaning:1 recently:3 common:9 endpoint:3 belong:5 he:2 interpretation:2 shenker:1 refer:7 rd:1 similarly:2 hp:2 had:1 moving:1 similarity:3 add:1 halved:1 recent:4 showed:1 perspective:2 inf:1 apart:1 reverse:3 certain:1 server:1 affiliation:2 continue:1 onr:1 molina:1 minimum:1 additional:1 somewhat:1 mr:1 goel:2 r0:1 determine:2 shortest:2 ii:4 full:2 technical:2 long:8 retrieval:1 visit:1 award:2 sigcomm:2 variant:1 basic:3 scalable:3 essentially:1 stant:1 represent:2 addition:1 want:1 fine:1 fellowship:1 else:1 median:1 ithaca:1 unlike:1 pass:1 induced:6 recruited:1 undirected:1 thing:1 december:1 balakrishnan:1 incorporates:1 seem:1 noting:1 easy:1 concerned:2 variety:2 affect:2 psychology:1 reduce:1 idea:2 prototype:1 itr:1 whether:1 six:3 motivated:1 url:1 linkage:2 passing:1 york:1 generally:1 outward:1 repeating:1 locally:1 morris:1 diameter:1 http:3 specifies:1 exist:3 nsf:2 notice:2 arising:1 per:2 ccr:1 group:43 key:1 killworth:2 blum:1 neither:1 ravi:1 graph:15 asymptotically:3 button:1 monotone:2 geometrically:1 fraction:2 inverse:1 letter:5 eli:1 striking:1 place:3 reasonable:2 reader:1 separation:2 home:1 duncan:2 clarke:1 bound:1 internet:2 followed:3 simplification:1 precisely:1 ri:3 kleinberg:4 kumar:1 department:1 designated:2 according:2 watt:7 ball:3 disconnected:1 belonging:2 across:4 slightly:1 increasingly:1 smaller:3 terminates:1 rev:1 den:1 indexing:1 resource:3 discus:2 r3:1 mechanism:1 know:5 available:2 decentralized:27 hierarchical:20 away:2 appropriate:1 original:1 recipient:2 gori:1 include:1 completed:1 tomkins:1 log2:2 music:2 giving:1 establish:1 society:1 seeking:1 move:1 implied:1 question:1 already:2 looked:1 strategy:2 exhibit:2 dp:1 distance:7 link:46 sci:2 topic:8 argue:1 extent:1 considers:1 trivial:1 toward:1 length:3 bollob:1 disruptive:1 october:1 statement:2 taxonomy:1 design:5 collective:1 unknown:2 observation:1 govindan:2 relational:2 extended:1 precise:1 communication:1 discovered:1 david:2 pair:9 required:1 concisely:1 polylogarithmic:23 nip:1 address:1 able:3 middleware:1 firstname:1 dimitris:1 hyperlink:1 rajagopalan:1 packard:1 analogue:2 power:3 getoor:1 natural:4 decidedly:1 scheme:1 improve:1 technology:1 discovery:2 occupation:4 law:2 girvan:1 fully:1 interesting:1 generation:1 proportional:5 limitation:1 infocom:1 funnel:1 foundation:2 upfal:1 norwood:1 agent:2 degree:14 gather:1 incident:1 principle:2 editor:1 share:1 collaboration:1 course:1 supported:1 free:1 allow:1 understand:2 wide:4 neighbor:4 taking:1 michelle:1 van:1 distributed:3 dimension:1 depth:3 world:15 superficially:1 forward:4 qualitatively:1 author:1 collection:6 far:2 social:7 obtains:1 abstracted:1 global:1 investigating:1 incoming:1 knew:1 search:41 nature:2 career:1 necessarily:1 protocol:4 motivation:1 whole:1 arise:1 sridhar:1 allowed:1 benefited:1 representative:1 gossip:3 ny:1 wiley:1 sub:10 learns:1 grained:1 young:1 theorem:11 specific:1 navigate:1 hub:1 r2:1 grouping:1 lukose:3 workshop:1 avrim:1 browsing:1 easier:1 logarithmic:1 garcia:1 simply:1 explore:1 likely:3 forming:3 hitting:1 contained:2 strogatz:5 partially:2 joined:1 corresponds:1 determines:1 acm:5 goal:4 viewed:1 banker:2 towards:1 kochen:1 kaashoek:1 content:2 broker:2 included:1 specifically:1 uniformly:3 huberman:3 operates:1 total:1 bernard:2 indicating:1 berg:1 internal:1 people:7 support:2 latter:1 mark:1 crawler:1 investigator:1 phenomenon:4 correlated:1
1,165
2,062
Estimating Car Insurance Premia: a Case Study in High-Dimensional Data Inference Nicolas Chapados, Yoshua Bengio, Pascal Vincent, Joumana Ghosn, Charles Dugas, Ichiro Takeuchi, Linyan Meng University of Montreal, dept. IRQ, CP 6128, Succ. Centre-Ville, Montreal, Qc, Canada, H3C3J7 {chapadosJbengioy,vincentp,ghosnJdugas,takeuchi,mengl}~iro.umontreal.ca Abstract Estimating insurance premia from data is a difficult regression problem for several reasons: the large number of variables, many of which are .discrete, and the very peculiar shape of the noise distribution, asymmetric with fat tails, with a large majority zeros and a few unreliable and very large values. We compare several machine learning methods for estimating insurance premia, and test them on a large data base of car insurance policies. We find that function approximation methods that do not optimize a squared loss, like Support Vector Machines regression, do not work well in this context. Compared methods include decision trees and generalized linear models. The best results are obtained with a mixture of experts, which better identifies the least and most risky contracts, and allows to reduce the median premium by charging more to the most risky customers. 1 Introduction The main mathematical problem faced by actuaries is that of estimating how much each insurance contract is expected to cost. This conditional expected claim amount is called the pure premium and it is the basis of the gross premium charged to the insured. This expected value is conditionned on information available about the insured and about the contract, which we call input profile here. This regression problem is difficult for several reasons: large number of examples, -large number variables (most of which are discrete and multi-valued), non-stationarity of the distribution, and a conditional distribution of the dependent variable which is very different from those usually encountered in typical applications .of machine learning and function approximation. This distribution has a mass at zero: the vast majority of the insurance contracts do not yield any claim. This distribution is also strongly asymmetric and it has fat tails (on one side only, corresponding to the large claims). In this paper we study and compare several learning algorithms along with methods traditionally used by actuaries for setting insurance premia. The study is performed on a large database of automobile insurance policies. The methods that were tried are the following: the constant (unconditional) predictor as a benchmark, linear regression, generalized linear models (McCullagh and NeIder, 1989), decision tree models (CHAID (Kass, 1980)), support vector machine regression (Vapnik, 1998), multi-layer neural networks, mixtures of neural network experts, and the current premium structure of the insurance company. In a variety of practical applications, we often find data distributions with an asymmetric heavy tail extending out towards more positive values. Modeling data with such an asymmetric heavy-tail distribution is essentially difficult because outliers, which are sampled from the tail of the distribution, have a strong influence on parameter estimation. When the distribution is symmetric (around the mean), the problems caused by outliers can be reduced using robust estimation techniques (Huber, 1982; F.R.Hampel et al., 1986; Rousseeuw and Leroy, 1987) which basically intend to ignore or downweight outliers. Note that these techniques do not work for an asymmetric distribution: most outliers are on the same side of the mean, so downweighting them introduces a strong bias on its estimation: the conditional expectation would be systematically underestimated. There is another statistical difficulty, due to the large number of variables (mostly discrete) and the fact that many interactions exist between them. Thus the traditional actuarial methods based on tabulating average claim amounts for combinations of values are quickly hurt by the curse of dimensionality, unless they make hurtful independence assumptions (Bailey and Simon, 1960). Finally, there is a computational difficulty: we had access to a large database of ~ 8 x 106 examples, and the training effort and numerical stability of some algorithms can be burdensome for such a large number of training examples. This paper is organized as follows: we start by describing the mathematical criteria underlying insurance premia estimation (section 2), followed by a brief review of the learning algorithms that we consider in this study, including our best-performing mixture of positive-output neural networks (section 3). We then highlight our most important experimental results (section 4), and in view of them conclude with an examination of the prospects for applying statistical learning algorithms to insurance modeling (section 5). 2 Mathematical Objectives The first goal of insurance premia modeling is to estimate the expected claim amount for a given insurance contract for a future one-year period (here we consider that the amount is 0 when no claim is filed). Let X E Rm denote the customer and contract input profile, a vector representing all the information known about the customer and the proposed insurance policy before the beginning of the contract. Let A E R+ denote the amount that the customer claims during the contract period; we shall assume that A is non-negative. Our objective is to estimate this claim amount, which is the pure premium Ppure of a given contract x: 1 Ppure(X) == E[AIX == x]. (1) The Precision Criterion. In practice, of course, we have no direct access to the quantity (1), which we must estimate. One possible criterion is to seek the most precise estimator, which minimizes the mean-squared error (MSE) over a data set D == {(xl,a?)}r=l. Let P == {p(?;8)} be a function class parametrized by the IThe pure premium is distinguished from the premium actually charged to the customer, which must account for the risk remaining with the insurer, the administrative overhead, desired profit, and other business costs. parameter vector (). The MSE criterion produces the most precise function (on average) within the class, as measured with respect to D: L ()* = argm:n ~ L(P(Xi; (}) - ai)2. (2) i=1 Is it an appropriate criterion and why? First one should note that if PI and P2 are two estimators of E[AIX]' then the MSE criterion is a good indication of how close they are to E[AIX], since by the law of iterated expectations, E[(PI(X) - A)2] - E[(P2(X) - A)2] == E[(PI(X) - E[AIX])2] -E[(P2(X) - E[AIX])2], and of course the expected MSE is minimized when p(X) == E[AIX]. The Fairness Criterion. However, in insurance policy pricing, the precision criterion is not the sole part of the picture; just as important is that the estimated premia do not systematically discriminate against specific segments of the population. We call this objective the fairness criterion. We define the bias of the premia b(P) to be the difference between the average premium and the average incurred amount, in a given population P: 1 (3) b(P) = 1FT p(Xi) - ai, L (xi,ai)EP where IPI denotes the cardinality of the set P, and p(.) is some premia estimation function. A possible fairness criterion would be based on minimizing the norm of the bias over every subpopulation Q of P. From a practical standpoint, such a minimization would be extremely difficult to carry out. Furthermore, the bias over small subpopulations is hard to estimate with statistical significance. We settle instead for an approximation that gives good empirical results. After training a model to minimize the MSE criterion (2), we define a finite number of disjoint subsets (subpopulations) of the test set P, PkC P, P k n Pj:f;k == 0, and verify that the absolute bias is not significantly different from zero. The subsets Pk can be chosen at convenience; in our experiments, we considered 10 subsets of equal-size delimited by the deciles of the test set premium distribution. In this way, we verify that, for example, for the group of contracts with a premium between the 5th and the 6th decile, the average premium matches the average claim amount. 3 Models Evaluated An important requirement for any model of insurance premia is that it should produce positive premia: the company does not want to charge negative money to its customers! To obtain positive outputs neural networks we have considered using an exponential activation function at the output layer but this created numerical difficulties (when the argument of the exponential is large, the gradient is huge). fustead, we have successfully used the "softplus" activation function (Dugas et al., 2001): softplus(s) == log(1 + e 8 ) where s is the weighted sum of an output neuron, and softplus(s) is the corresponding predicted premium. Note that this function is convex, monotone increasing, and can be considered as a smooth version of the "positive part" function max(O, x). The best model that we obtained is a mixture of experts in which the experts are positive outputs neural networks. The gater network (Jacobs et al., 1991) has softmax outputs to obtain positive w~ights summing to one. X 10-3 Distribution of (claim - prediction) in each prediction quintile 2 1.8 1.6 1.4 1.2 0.8 0.6 0.4 0.2 oL-.._..I....=~-L-~-l...-_----L_----l-==:::=:~::=::::::r:::===:?==~ -3000 -2000 -1000 1000 2000 claim - prediction 3000 4000 5000 6000 Proportion of non-zero claims in each prediction quintile 0.25 r - - - r - - - - - - - , r - - - - - - - - - - , - - - - - - , . - - - - - - - - , - - - , 0.15 0.1 0.05 3 quintile Figure 1: A view of the conditional distribution of the claim amounts in the out-ofsample test set. Top: probability density of (claim amount - conditional expectation) for 5 quintiles of the conditional expectation, excluding zero-claim records. The mode moves left for increasing conditional expectation quintiles. Bottom: proportion of non-zero claim records per quintile of the prediction. The mixture model was compared to other models. The constant model only has intercepts as free parameters. The linear model corresponds to a ridge linear regression (with weight decay chosen with the validation set). Generalized linear models (GLM) estimate the conditional expectation from j(x) == eb+w1x with parameters b and w. Again weight decay is used and tuned on the validation set. There are many variants of GLMs and they are popular for building insurance models, since they provide positive outputs, interpretable parameters, and can be associated to parametric models of the noise. Decision trees are also used by practitioners in the insurance industry, in particular the CHAID-type models (Kass, 1980; Biggs, Ville and Suen, 1991), which use statistical criteria for deciding how to split nodes and when to stop growing the tree. We have compared our models with a CHAID implementation based on (Biggs, Ville and Suen, 1991), adapted for regression purposes using a MANOVA analysis. The threshold parameters were selected based on validation set MSE. Regression Support Vector Machines (SVM) (Vapnik, 1998) were also evaluated Mean-Squared Error 67.1192 .......................................................................................... :;....:-.:--'* ..... 67.0851 .... -*"------ ~~--:-~-:-~~: :.~:-.:- - -~-~--:.- -.~~--'---~ : .. .. . . -.. . . _.. .. Test . 56.5744 Validation 56.5416 56.1108 . Training 56.0743 Figure 2: MSE results for eight models. Models have been sorted in ascending order of test results. The training, validation and test curves have been shifted closer together for visualization purposes (the significant differences in MSE between the 3 sets are due to "outliers"). The out-of-sample test performance of the Mixture model is significantly better than any of the other. Validation based model selection is confirmed on test results. CondMean is a constructive greedy version of GLM. but yielded disastrous results for two reasons: (1) SVM regression optimizes an L 1 like criterion that finds a solution close to the conditional median, whereas the MSE criterion is minimized for the conditional mean, and because the distribution is highly asymmetric the conditional median is far from the conditional mean; (2) because the output variable is difficult to predict, the required number of support vectors is huge, also yielding poor generalization. Since the median is actually 0 for our data, we tried to train the SVM using only the cases with positive claim amounts, and compared the performance to that obtained with the GLM and the neural network. The SVM is still way off the mark because of the above two reasons. Figure 1 (top) illustrates the fat tails and asymetry of the conditional distribution of the claim amounts. . Finally, we compared the best statistical model with a proprietary table-based and rule-based premium estimation method that was provided to us as the benchmark against which to judge improvements. 4 Experimental Results Data from five kinds of losses were included in the study (Le. a sub-premium was estimated for each type of loss), but we report mostly aggregated results showing the error on the total estimated premium. The input variables contain information about the policy (e.g., the date to deal with inflation, deductibles and options), the car, and the driver (e.g., about past claims, past infractions, etc...). Most variables are subject to discretization and binning. Whenever possible, the bins are chosen such that they contain approximately the same number of observations. For most models except CHAID, the discrete variables are one-hot encoded. The number of input random variables is 39, all discrete except one, but using one-hot encoding this results in an input vector x of length m == 266. An overall data set containing about Table 1: Statistical comparison of the prediction accuracy difference between several individual learning models and the best Mixture model. The p-value is given under the null hypothesis oino difference between Model #1 and the best Mixture model. Note that all differences are statistically significant. Model #1 Model #2 Constant Mixture CHAID Mixture GLM Mixture Softplus NN Mixture Linear Mixture Mixture NN Mean MSE Diff. 3.40709e-02 2.35891e-02 7.54013e-03 6.71066e-03 5.82350e-03 5.23885e-03 Std. Error 3.32724e-03 2.57762e-03 1.15020e-03 1.09351e-03 1.32211e-03 1.41112e-03 Z p-value 10.2400 9.1515 6.5555 6.1368 4.4047 3.7125 0 0 2.77e-ll 4.21e-l0 5.30e-06 1.02e-04 Table 2: MSE difference between benchmark and Mixture models across the 5 claim categories (kinds of losses) and the total claim amount. In all cases except category 1, the IvIixture model is statistically significantly (p < 0.05) more precise than the benchmark model. Claim Category (Kind of Loss) Category 1 Category 2 Category 3 Category 4 Category 5 Total claim amount MSE Difference Benchmark minus Mixture 20669.53 1305.57 244.34 1057.51 1324.31 60187.60 95% Confidence Interval Lower Higher (-4682.83 - 46021.89 ) (1032.76 1578.37 ) (6.12 482.55 ) (623.42 1491.60 ) (1077.95 1570.67 ) ( 7743.96 - 112631.24) 8 million examples is randomly permuted and split into a training set, validation set and test set, respectively of size 50%, 25% and 25% of the total. The validation set is used to select among models (includi~g the choice of capacity), and th~ test set is used for final statistical comparisons. Sample-wise paired statistical tests are used to reduce the effect of huge per-sample variability. Figure 1 is an attempt at capturing the shape of the conditional distribution of claim amounts given input profiles, by considering the distributions of claim amounts in different quantiles of the prediction (pure premium), on the test set. The top figure excludes the point mass of zero claims and rather shows the difference between the claim amount and the estimated conditional expectation (obtained with the mixture model). The bottom histogram shows that the fraction of claims increases nicely for the higher predicted pure premia. Table 1 and Figure 2 summarize the comparison between the test MSE of the different tested models. NN is a neural network with linear output activation whereas Softplus NNhas the softplus output activations. The Mixture is the mixture of softplus neural networks. This result identifies the mixture model with softplus neural networks as the best-performing of the tested statistical models. Our conjecture is that the mixture model works better because it is more robust to the effect of "outliers" (large claims). Classical robust regression methods (Rousseeuw and Leroy, 1987) work by discarding or downweighting outliers: they cannot be applied here because the claims distribution is highly asymmetric (the extreme values are always large ones, the claims being all non-negative). Note that the capacity of each model has been tuned on the validation set. Hence, e.g. CHAID could have easily yielded lower training error, but at the price of worse generalization. x10 4 Rule-Based minus UdeM Mixture 2,...-------,------,-----r-------.-------,-----.---------.----......., Mean = -1.5993e-1 a Median = 37.5455 ... Stddev = 154.65 - 1.5 ~ o c(]) :::l 0(]) u: 0.5 OL.-..----L.----L----.L.----..L---~~ -3000 -2500 -2000 -1500 -1000 -500 Difference between premia ($) o 500 1000 Figure 3: The premia difference distribution is negatively skewed, but has a positive m~dian for a mean of zero. This implies that the benchmark model (current pricing) undercharges risky customers, while overcharging typical customers. Table 2 shows a comparison of this model against the rule-based benchmark. The improvements are shown across the five types of losses. In all cases the mixture improves, and the improvement is significant in four out of the five as well as across the sum of the five. A qualitative analysis of the resulting predicted premia shows that the mixture model has smoother and more spread-out premia than the benchmark. The analysis (figure 3) also reveals that the difference between the mixture premia and the benchmark premia is negatively skewed, with a positive median, i.e., the typical customer will pay less under the new mixture model, but the "bad" (risky) customers will pay much more. To evaluate fairness, as discussed in the previous section, the distribution of premia computed by the best model is analyzed, splitting the contracts in 10 groups according to their premium level. Figure 4 shows that the premia charged are fair for each sub-population. 5 Conclusion This paper illustrates a successful data-mining application in the insurance industry. It shows that a specialized model (the mixture model), that was designed taking into consideration the specific problem posed by the data (outliers, asymmetric distribution, positive outputs), performs significantly better than existing and popular learning algorithms. It also shows that such models can significantly improve over the current practice, allowing to compute premia that are lower for less risky contracts and higher for more risky contracts, thereby reducing the cost of the median contract. Future work should investigate in more detail the role of temporal pon-stationarity, how to optimize fairness (rather than just test for it afterwards), and how to further increase the robustness of the model with respect to large claim amounts. Difference with incurred claims (sum of all KOL-groups) 200 ............ 0 ? .. .. . 0 0 .' .. .. ~ C/} E ~ 0 "'C ~ :5 "~ -200 -5 I :? : :1' : ?1????????? .\. . \ . . . . .......:? ? \ \ "? CD ~ -400 CD ~ o -600 -B- Mixture Model (normalized premia) -* - Rule-Based Model (normalized premia) 2 4 6 8 10 Decile Figure 4: We ensure fairness by comparing the average incurred amount and premia within each decile of the premia distribution; both models are generally fair to subpopu1ations. The error bars denote 95% confidence intervals. The comparisqn is for the sum of claim amounts over all 5 kinds of losses (KOL). References Bailey, R. A. and Simon, L. (1960). Two studies in automobile insurance ratemaking. ASTIN Bulletin, 1(4):192-217. Biggs, D., Ville, B., and Suen, E. (1991). A method of choosing multiway partitions for classification and decision trees. Journal of Applied Statistics, 18(1):49-62. Dugas, C., Bengio, Y., Belisle, F., and Nadeau, C. (2001). Incorporating second order functional? knowledge into learning algorithms. In Leen, T., Dietterich, T., and Tresp, V., editors, Advances in Neural Information Processing Systems, volume 13, pages 472-478. F.R.Hampel, E.M.Ronchetti, P.J.Rousseeuw, and W.A.Stahel (1986). Robust Statistics, The Approach based on Influence Functions. John Wiley & Sons. Huber, P. (1982). Robust Statistics. John Wiley & Sons Inc. Jacobs, R. A., Jordan, M. I., Nowlan, S. J., and Hinton, G. E. (1991). Adaptive mixture of local experts. Neural Computation, 3:79-87. Kass, G. (1980). An exploratory technique for investigating large quantities of categorical data. Applied Statistics, 29(2):119-127. McCullagh, P. and NeIder, J. (1989). Generalized Linear Models. Chapman and Hall, London. ' Rousseeuw, P. and Leroy, A. (1987). Robust Regression and Outlier Detection. John Wiley & Sons Inc. Vapnik, V. (1998). Statistical Learning Theory. Wiley, Lecture Notes in Economics and Mathematical Systems, volume 454.
2062 |@word version:2 norm:1 proportion:2 seek:1 tried:2 jacob:2 ronchetti:1 thereby:1 profit:1 minus:2 carry:1 tuned:2 past:2 existing:1 ka:3 current:3 discretization:1 comparing:1 nowlan:1 activation:4 must:2 john:3 numerical:2 partition:1 shape:2 designed:1 interpretable:1 greedy:1 selected:1 beginning:1 argm:1 stahel:1 record:2 node:1 ipi:1 five:4 mathematical:4 along:1 direct:1 driver:1 qualitative:1 overhead:1 huber:2 expected:5 growing:1 multi:2 ol:2 company:2 curse:1 cardinality:1 increasing:2 considering:1 provided:1 estimating:4 underlying:1 mass:2 null:1 kind:4 minimizes:1 temporal:1 every:1 charge:1 fat:3 rm:1 decile:4 positive:12 before:1 local:1 encoding:1 meng:1 approximately:1 eb:1 statistically:2 practical:2 practice:2 gater:1 empirical:1 significantly:5 confidence:2 subpopulation:3 cannot:1 close:2 selection:1 convenience:1 context:1 influence:2 applying:1 risk:1 intercept:1 optimize:2 customer:10 charged:3 pon:1 economics:1 convex:1 qc:1 splitting:1 pure:5 estimator:2 rule:4 stability:1 population:3 exploratory:1 traditionally:1 hurt:1 hypothesis:1 asymmetric:8 std:1 database:2 binning:1 ep:1 ft:1 bottom:2 role:1 prospect:1 gross:1 segment:1 ithe:1 negatively:2 basis:1 biggs:3 easily:1 succ:1 train:1 actuarial:1 london:1 choosing:1 quintiles:2 encoded:1 posed:1 valued:1 statistic:4 final:1 indication:1 dian:1 interaction:1 date:1 requirement:1 extending:1 produce:2 montreal:2 measured:1 sole:1 p2:3 strong:2 predicted:3 judge:1 implies:1 settle:1 bin:1 generalization:2 around:1 considered:3 inflation:1 hall:1 deciding:1 predict:1 claim:34 purpose:2 estimation:6 successfully:1 weighted:1 minimization:1 suen:3 always:1 rather:2 l0:1 improvement:3 burdensome:1 inference:1 dependent:1 nn:3 overall:1 among:1 classification:1 pascal:1 softmax:1 equal:1 nicely:1 chapman:1 fairness:6 future:2 minimized:2 yoshua:1 report:1 few:1 randomly:1 individual:1 attempt:1 detection:1 stationarity:2 huge:3 highly:2 mining:1 investigate:1 insurance:20 w1x:1 introduces:1 mixture:29 extreme:1 analyzed:1 yielding:1 unconditional:1 peculiar:1 closer:1 unless:1 tree:5 desired:1 industry:2 modeling:3 insured:2 cost:3 subset:3 predictor:1 successful:1 quintile:4 density:1 filed:1 contract:14 off:1 aix:6 together:1 quickly:1 squared:3 again:1 containing:1 actuary:2 worse:1 expert:5 account:1 inc:2 caused:1 performed:1 view:2 neider:2 ichiro:1 start:1 option:1 astin:1 simon:2 minimize:1 takeuchi:2 accuracy:1 chapados:1 yield:1 vincent:1 iterated:1 basically:1 confirmed:1 whenever:1 against:3 associated:1 sampled:1 stop:1 popular:2 knowledge:1 car:3 dimensionality:1 improves:1 organized:1 actually:2 delimited:1 higher:3 leen:1 evaluated:2 strongly:1 furthermore:1 just:2 glms:1 ghosn:1 mode:1 pricing:2 tabulating:1 building:1 effect:2 dietterich:1 verify:2 contain:2 normalized:2 hence:1 symmetric:1 deal:1 ll:1 during:1 skewed:2 criterion:14 generalized:4 ridge:1 performs:1 cp:1 wise:1 consideration:1 charles:1 umontreal:1 specialized:1 permuted:1 functional:1 volume:2 million:1 tail:6 discussed:1 significant:3 ai:3 centre:1 multiway:1 had:1 access:2 money:1 etc:1 base:1 pkc:1 optimizes:1 irq:1 aggregated:1 period:2 smoother:1 afterwards:1 x10:1 smooth:1 match:1 dept:1 paired:1 prediction:7 variant:1 regression:11 essentially:1 expectation:7 histogram:1 whereas:2 want:1 interval:2 underestimated:1 median:7 standpoint:1 subject:1 jordan:1 call:2 practitioner:1 bengio:2 split:2 variety:1 independence:1 reduce:2 effort:1 proprietary:1 generally:1 amount:20 rousseeuw:4 category:8 reduced:1 exist:1 shifted:1 estimated:4 disjoint:1 per:2 discrete:5 shall:1 group:3 four:1 threshold:1 downweighting:2 pj:1 kol:2 vast:1 excludes:1 ville:4 monotone:1 year:1 sum:4 fraction:1 decision:4 capturing:1 layer:2 pay:2 followed:1 encountered:1 yielded:2 leroy:3 adapted:1 argument:1 extremely:1 performing:2 downweight:1 conjecture:1 according:1 combination:1 poor:1 across:3 son:3 outlier:9 glm:4 visualization:1 describing:1 ascending:1 available:1 eight:1 appropriate:1 bailey:2 distinguished:1 robustness:1 denotes:1 remaining:1 include:1 top:3 ensure:1 classical:1 objective:3 intend:1 move:1 quantity:2 parametric:1 traditional:1 gradient:1 capacity:2 majority:2 parametrized:1 iro:1 reason:4 length:1 minimizing:1 difficult:5 mostly:2 disastrous:1 negative:3 implementation:1 policy:5 allowing:1 neuron:1 observation:1 hurtful:1 benchmark:9 finite:1 dugas:3 hinton:1 excluding:1 precise:3 variability:1 canada:1 required:1 bar:1 usually:1 ofsample:1 vincentp:1 summarize:1 including:1 max:1 charging:1 hot:2 difficulty:3 examination:1 hampel:2 business:1 representing:1 improve:1 brief:1 risky:6 picture:1 identifies:2 created:1 nadeau:1 categorical:1 tresp:1 faced:1 review:1 law:1 loss:7 lecture:1 highlight:1 validation:9 incurred:3 editor:1 systematically:2 pi:3 heavy:2 cd:2 course:2 free:1 l_:1 side:2 bias:5 taking:1 bulletin:1 absolute:1 curve:1 adaptive:1 ights:1 far:1 belisle:1 ignore:1 unreliable:1 reveals:1 investigating:1 summing:1 conclude:1 xi:3 why:1 table:5 robust:6 nicolas:1 ca:1 mse:13 automobile:2 significance:1 main:1 pk:1 spread:1 noise:2 profile:3 fair:2 quantiles:1 wiley:4 precision:2 sub:2 exponential:2 xl:1 administrative:1 bad:1 specific:2 discarding:1 showing:1 decay:2 svm:4 incorporating:1 vapnik:3 illustrates:2 manova:1 corresponds:1 conditional:15 goal:1 sorted:1 towards:1 price:1 hard:1 mccullagh:2 included:1 typical:3 except:3 diff:1 reducing:1 called:1 total:4 discriminate:1 experimental:2 premium:17 select:1 support:4 softplus:8 mark:1 constructive:1 evaluate:1 tested:2
1,166
2,063
Thomas L . Griffiths & Joshua B. Tenenbaum Department of Psychology Stanford University, Stanford, CA 94305 {gruffydd,jbt}?psych. stanford. edu Abstract Estimating the parameters of sparse multinomial distributions is an important component of many statistical learning tasks. Recent approaches have used uncertainty over the vocabulary of symbols in a multinomial distribution as a means of accounting for sparsity. We present a Bayesian approach that allows weak prior knowledge, in the form of a small set of approximate candidate vocabularies, to be used to dramatically improve the resulting estimates. We demonstrate these improvements in applications to text compression and estimating distributions over words in newsgroup data. 1 Introduction Sparse multinomial distributions arise in many statistical domains, including natural language processing and graphical models. Consequently, a number of approaches to parameter estimation for sparse multinomial distributions have been suggested [3]. These approaches tend to be domain-independent: they make little use of prior knowledge about a specific domain. In many domains where multinomial distributions are estimated there is often at least weak prior knowledge about' the potential structure of distributions, such as a set of hypotheses about restricted vocabularies from which the symbols might be generated. Such knowledge can be solicited from experts or obtained from unlabeled data. We present a method for Bayesian_parameter estimation in sparse discrete domains that exploits this weak form of prior knowledge to improve estimates over knowledge-free approaches. 1.1 Bayesian parameter estimation for multinomial distributions Following the presentation in [4], we consider a language ~ containing L distinct symbols. A multinomial distribution is specified by a parameter vector f) == (Ol, ... ,f)L), where f)i is the probability of an observation being symbol i. Consequently, we have the constraints that Ef==l f)i == 1 and (h ~ 0, i == 1, ... ,L. The task of multinomial estimation is to take a data set D and produce a'vector f) that results in a good approximation to the distribution that produced D. In this case, D consists of N independent observations Xl, ... x N drawn from the distribution to be estimated, which can be summarized by the statistics N i specifying the number of times the ith symbol occurs in the data. D also determines the set ~o of symbols that occur in the data. Stated in this way, multinomial estimation involves predicting the next observation based on the data. Specifically, we wish to calculate P(XN+1ID). The Bayesian estimate for this probability is given by PL(xN+lID) = I p(XN+1IB)P(BID)dB where P(X N + 1 10) follows from the multinomial distribution corresponding to O. The posterior probability P(OID) can be obtained via Bayes rule L P(OID) oc P(DIO)P(O) == P(8} II ONi i==l where P(O) is the prior probability of a given O. Laplace used this method with a uniform prior over 0 to give the famous "law of succession" [6J. A more general approach is to assume a Dirichlet prior over (), which is conjugate to the multinomial distribution and gives N i +LCY.i P(XN+l = ilD) = N (1) + l:j==l O!.j where the ai are the hyperparameters of the Dirichlet distribution. Different estimates are obtained for different choices of the ai, with most approaches making the simplifying assumption that ai == O!. for all i. Laplace's law results from a == 1. The case with a == 0.5 is the Jeffreys-Perks law or Expected Likelihood Estimation [2] [5J [9J, while using arbitrary O!. is Lidstone's law [7]. 1.2 EstiIllating sparse Illultinomial distributions Several authors have extended the Bayesian approach to sparse multinomial distributions, in which only a restricted vocabulary of symbols are used, by maintaining uncertainty over these vocabularies. In [10], Ristad uses assumptions about the probability of strings based upon different vocabularies to give the estimate PR (X N +1 == ilD) == (Ni + l)/(N +L) (Ni + l)(N + 1 - kO)/(N 2 + N + 2kO) { kO(kO + l)/(L - kO)(N 2 + N + 2kO) if kO == L if kO < L 1\ N i otherwise >0 where kO == I~o I is the size of the smallest vocabulary consistent with the data. A different approach is taken by Friedman and Singer in [4], who point out that Ristad's method is a special case of their framework. Friedman and Singer consider the vocabulary V ~ :E to be a random variable, allowing them to write p.(X N +1 == ilD) == L p(X N +1 == ilV, D)P(VID) (2) v where P{X N + 1 == ilV, D) results from a Dirichlet prior over the symbols in V, p(X N +1 == ilV, D) == {~it'ja o v: if i E otherWIse (3) and by Bayes' rule and the properties of Dirichlet priors P(VID) oc P(DIV)P(V) { ~fJ~~(a) niE~O r(~?t~a) P(V) EO ~ V otherwise ( ) 4 Friedman and Singer assume a hierarchical prior over V, such that all vocabularies of cardinality k are given equal probability, namely P(S == k)/(t), where P(S == k) is the probability that the size of the vocabulary (IVI) is k. It follows that if i E ~o, p(X N + I == ilD) == Lk :+1~P(S == kiD). If i ? ~o, it is necessary to estimate the proportion of V that contain i for a given k. The simplified result is PF(X N +1 == ilD) == { %tt~aC(D,L) L-k O (1- C(D,L)) if i E ~o otherwise (5) where .h WIt 2 mk P(S == k) (k-kO)!? - k! ==. r(ka:) r(N+ka:) . Ivlaking use of weak prior knowiedge Friedman and Singer assume a prior that gives equal probability to all vocabularies of a given cardinality. However, many real-world tasks provide limited knowledge about the structure of distributions that we can build into our methods for parameter estimation. In the context of sparse multinomial estimation, one instance of such knowledge the importance of specific vocabularies. For example, in predicting the next character in a file, our predictions could be facilitated by considering the fact that most files either use a vocabulary consisting of ASCII printing characters (such as text files), or all possible characters (suc~ as object files). This kind of structural knowledge about a domain is typically easier to solicit from experts than accurate distributional information, and forms a valuable informational resource. If we have this kind of prior knowledge, we can restrict our attention to a subset of the 2L possible vocabularies. fu particular, we can specify a set of vocabularies V which we consider as hypotheses for the vocabulary used in producing D, where the elements of V are specified by our knowledge of the domain. This stands as a compromise between Friedman and Singer's approach, in which V consists of all vocabularies, and traditional Bayesian parameter estimation as represented by Equation 1, in which V consists of only the vocabulary containing all words. To do this, we explicitly evaluate the sum given in Equation 2, where the sum over V includes all V E V. This sum remains tractable when V is a small subset of the possible vocabularies, and the efficiency is aided by the fact that P(DIV) shares common terms across all V which can cancel in normalization. The intuition behind this approach is that it attempts to classify the target distribution as using one of a known set of vocabularies, where the vocabularies are obtained either from experts or from unlabeled data. Applying standard Bayesian multinomial estimation within this vocabulary gives enough flexibility for the method to capture a range of distributions, while making use of our weak prior knowledge. 2.1 An illustration: Text compression Text compression is an effective test of methods for multinomial estimation. Adaptive coding can be performed by specifying a method for calculating a distribution over the probability of the next byte in a file based upon the preceding bytes [1]. The extent to which the file is compressed depends upon the quality of these predictions. To illustrate the utility of including prior knowledge, we follow Ristad in using the Calgary text compression corpus [1]. This corpus consists of 19 files of Table 1: Text compression lengths (in bytes) on the Calgary corpus file bib book1 book2 geo nellS obj1 obj2 paper1 paper2 paper3 paper4 paper5 paper6 pic progc progl progp trans size kO NH(Ni/ N ) Pv PF PR PL PJ 111261 768771 610856 102400 377109 21504 246814 53161 82199 46526 13286 11954 38105 513216 39611 71646 49379 93695 81 82 96 256 98 256 256 95 91 84 80 91 93 159 92 87 89 99 72330 435043 365952 72274 244633 15989 193144 33113 47280 27132 7806 7376 23861 77636 25743 42720 30052 64800 18 219 94 161 89 126 182 71 75 70 58 57 68 205 68 74 71 169 89 105 115 162 113 127 184 94 94 85 72 79 90 92 116 124 165 116 129 190 100 105 92 79 83 95 216 91 97 94 105 269 352 329 165 304 129 189 236 259 238 190 181 223 323 222 253 236 252 174 219 212 161 201 126 182 156 167 154 126 122 149 205 150 164 155 169 16~ - 89 91 89 101 several different types, each using some subset of 256 possible characters (L == 256). The files include Bib'IEXsource (bib), formatted English text (book*, paper*), geological data (geo), newsgroup articles (news), object files (obj*), a bit-mapped picture (pic), programs in three different languages (prog*) and a terminal transcript (trans). The task was to estimate the distribution from which characters in the file were drawn based upon the first N characters and thus predict the N + 1st character. Performance was measured in terms of the length of the resulting file, where the contribution of the N + 1st character to the length is log2 P(XN+lID). The results are expressed as the number of bytes required to encode the file relative to the empirical entropy NH(Ni/N) as assessed by Ristad [10]. Results are shown in Table 1. P v is the restricted vocabulary model outlined above, with V consisting of just two hypotheses: one corresponding to binary files, containing all 256 characters, and one consisting of a 107 character vocabulary representing formatted English. The latter vocabulary was estimated from 5MB of English text, C code, Bib'IEXsource, and newsgroup data from outside the Calgary corpus. PF is Friedman and Singer's method. For both of these approaches, a was set to 0.5, to allow direct comparison to the Jeffreys-Perks law, PJo PR and PL are Ristad's and Laplace's laws respectively. P y outperformed the other methods on all files based upon English text, bar bookl, and all files using all 256 symbols l . The high performance followed from rapid classification of these files as using the appropriate vocabulary in V. When the vocabulary included all symbols Py performed as PJ, which gave the best predictions for these files. 1 A number of excellent techniques for? text compression exist that outperform all of those presented here. We have not included these techniques for comparison because our interest is in using text compression as a means of assessing estimation procedures, rather than as an end in itself. We thus consider only methods for multinomial estimation as our comparison. group. 2.2 Maintaining uncertainty in vocabularies The results for book1 illustrate a weakness of the approach outlined above. The file length for P y is higher than those for PF and PR , despite the fact that the file uses a text-based vocabulary. This file contains two characters that were not encountered in the data used to construct V. These characters caused P y to default to the unrestricted vocabulary of all 256 characters. From that point P y corresponded to PJ, which gave poor results on this file. This behavior results from the assumption that the candidate vocabularies in V are completely accurate. Since in many cases the knowledge that informs the vocabularies in V may be imperfect, it is desirable to allow for uncertainty in vocabularies. This uncertainty will be reflected in the fact that symbols outside V are expected to occur with a vocabulary-specific probability ty, p(XN+1 == ilV, D) == { (1 - (L -IVI)ty) N~~t~la ty where Ny == I:iEY P(DIV) if i E V otherwise N i ? It follows that = (1 - r(Ni + a) r(a:) r(IVla) + 1V1a:) v (L -IVJ)?V)NV ?t"-N r(N y which replaces Equations 3-4 in specifying P y iE:EonY . When V is determined by the judgments of domain experts, ty is the probability that an unmentioned word actually belongs to a particular vocabulary. While it may not be the most efficient use of such data, the V E V can also be estimated from some form of unlabeled data. In this case, Friedman and Singer's approach provides a means of setting ty. Friedman and Singer explicitly calculate the probability that an unseen word is in V based upon a dataset: from the second condition of Equation 5, we find that we should set ty == L_1IYI (1- C(D, L)). We use this method below. 3 Bayesian parameter estimation in natural language Statistical natural language processing often uses sparse multinomial distributions over large vocabularies of words. In different contexts, different vocabularies will be used. By specifying a basis set of vocabularies, we can perform parameter estimation by classifying distributions according to their vocabulary. This idea was examined using data from 20 different Usenet newsgroups. This dataset is commonly used in testing text classification algorithms (eg. [8]). Ten newsgroups were used to estimate a set of vocabularies V with corresponding ty. These vocabularies were used in estimating multinomial distributions on these newsgroups and ten others. The dataset was 20news-18827, which consists of the 20newsgroups data with headers and duplicates removed, and was preprocessed to remove all punctuation, capitalization, and distinct numbers. The articl~s in each of the 20 newsgroups were then divided into three sets. The first 500 articles from ten newsgroups were used to estimate the candidate vocabularies V and uncertainty parameters ty. Articles 501700 for all 20 newsgroups were used as training data for multinomial estimation. Articles 701-900 for all 20 newgroups were used as testing data. Following [8], a dictionary was built up by running over the 13,000 articles resulting from this division, and all words that occurred only once were mapped to an "unknown" word. The resulting dictionary contained L == 54309 words. As before, the restricted vocabulary method (Py), Friedman and Singer's method (PF ), and Ristad's (PR ), Laplace's (PL ) and the Jeffreys-Perks (PJ ) laws were ap- alt.atheism talk.politics.mideast 18 .... 11 " talk.politics. mise - 18 . 17 '. 16 scLspace rec.motorcycles ~~~ ~~~ ~~~ ~~~ soc.religion.christian talk.politics.guns comp.sys.ibm.pc.hardware rec.sport.hockey scLelectronics comp.windows.x rec.autos rec.sport.baseball scLcrypt scLmed comp.os.ms-windows.misc misc.forsale ~::=:= ~-;""", r.?';'~;:;" ~~~ comp.sys.mac.hardware 100 10000 Number of words 50000 talk.religion.misc comp.graphics ~~~ 100 10000 F~gure 1: Cross-entropy of predictions on newsgroup data as a function of the logarithm of the number of words. The abscissa is at the empirical entropy of the test distribution. The top ten panels (talk.polities.mideast and those to its right) are for the newsgroups with unknown vocabularies. The bottom ten are for those that contributed vocabularies to V, trained and tested on novel data. PL and P J are both indicated with dotted lines, but P J always performs better than PL. The box on talk.polities.mideast indicates the point at which Pv defaults to the full vocabulary, as the number of unseen words makes this vocabulary more likely. At this point, the line for Pv joins the line for P J , since both methods give the same estimates of the distribution. plied to the task. Both P v and PF used a == 0.5 to facilitate comparison with P J . 'V featured one vocabulary that contained all words in the dictionary, and ten vocabularies each corresponding to the words used in the first 500 articles of one of the newsgroups designated for this purpose. ?y was estimated as outlined above. Testing for each newsgroup consisted of taking words from the 200 articles assigned for training purposes, estimating a. distribution using each method, and then computing the cross-entropy between that distribution and an empirical estimate of the true distribution. The cross-entropy is H{Q; P) == Ei Qi log2 Pi, where Q is the true distribution and P is the distribution produced by the estimation method. Q was given by the maximum likelihood estimate formed from the word frequencies in all 200 articles assigned for testing purposes. The testing procedure was conducted with just 100 words, and then in increments of 450 up to a total of 10000 words. Long-run performance was examined on talk.polities.mideast and talk.polities.mise, each trained with 50000 words. The results are shown in Figure 1. As expected, P y consistently outperformed the other methods on the newsgroups that contributed to V. However, performance on novel newsgroups was also greatly improved. As can be seen in Figure 2, the novel newsgroups were classified to appropriate vocabularies - for example all words rec.autos I-----------------rec.motorcycles rec.sport.baseball scLcrypt scLmed ta1k.politics.guns talk.politics.mideast r l \ - - _ f - - - - - - - T - - - - ' r - - - - - - - - - - - alt.atheism f r t - ' \ . ; : : : : : : : : : : : ; f : = . = ' = f - - t - - - - - - - - - - soc.religion.christian r--+-- talk. politics. misc ,-------T---+----------- talk. religion. misc l-+-if-Hf-------t---------- ,-- misc.forsale scLspace ~~?.;~~~~~~g~ey comp.sys.mac.hardware comp.os.ms-windows.m~c'-----------------com~sy&ibm.p~hardware comp.graphics o 10000 Number of words Figure 2: Classification of newsgroup vocabularies. The lines illustrate the vocabulary which had maximum posterior probability for each of the ten test newsgroups after exposure to differing numbers of words. The vocabularies in V are listed along the left hand side of the axis, and the lines are identified with newsgroups by the labels on the right hand side. Lines are offset to facilitate identification. talk.religion.misc had the highest posterior probability for alt.atheism and soc. religion. christian, while rec. autos had highest posterior probability for rec .motorcycles. The proportion of word types occurring in the test data but not the vocabulary to which the novel newsgroups were classified ranged between 30.5% and 66.2%, with a mean of 42.2%. This illustrates that even approximate knowledge can facilitate predictions: the basis set of vocabularies allowed the high frequency words in the data to be modelled effectively, without excess mass being attributed to the low frequency novel word tokens. Long-run performance on talk.politics .mideast illustrates the same defaulting behavior that was shown for text classification: when the data become more probable under the vocabulary containing all words than under a restricted vocabulary the method defaults to the Jeffreys-Perks law. This guarantees that the method will tend to perform no worse than P J when unseen words are encountered in sufficient proportions. This is desirable, since PJ gives good estimates once N becomes large. 4 Discussion Bayesian approaches to parameter estimation for sparse multinomial distributions have employed the notion of a restricted vocabulary from which symbols are produced. In many domains where such distributions are estimated; there is often at least limited knowledge about the structure of these vocabularies. By considering just the vocabularies suggested by such knowledge, together with some uncertainty concerning those vocabularies, we can achieve very good estimates of distributions in these domains. We have presented a Bayesian approach that employs limited prior knowledge, and shown that it outperforms a range of approaches to multinomial estimation for both text compression and a task involving natural language. While our applications in this paper estimated approximate vocabularies from data, the real promise of this approach lies with domain knowledge solicited from experts. Experts are typically better at providing qualitative structural information than quantitative distributional information, and our approach provides a way of using this information in parameter estimation. For example, in the context of parameter estimation for graphical models to be used in medical diagnosis, distinguishing classes of symptoms might be informative in determining the parameters governing their relationship to diseases. This form of knowledge is naturally translated into a set of vocabularies to be considered for each such distribution. More complex applications to natural language lllay also be possible, such as using syntactic information in estimating probabilities for n-gram models. The approach we have presented in this paper provides a simple way to allow this kind of limited domain knowledge to be useful in Bayesian parameter estimation. References [1] T. C. Bell, J. G. Cleary, and 1. H. Witten. Text compression. Prentice Hall, 1990. [2] G. E. P. Box and G. C. Tiao. Bayesian Inference in Statistical Analysis. AddisonWesley, 1973. [3] S. F. Chen and J. Goodman. An empirical study of smoothing techniques for language modeling. Technical Report TR-10-98, Center for Research in Computing Technology, Harvard University, 1998. [4] N. Friedman and Y. Singer. Efficient Bayesian parameter estimation in large discrete domains. In Neural Information Processing Systems, 1998. [5] H. Jeffreys. An invariant form for the prior probability in estimation problems. Proceedings of the Royal Society A, 186:453-461, 1946. [6] P.-S. Laplace. Philosophical Essay on Probabilities. Springer-Verlag, 1995. Originally published 1825. [7] G. Lidstone. Note on the general case of the Bayes-Laplace formula for inductive or a posteriori probabilities. Transactions of the Faculty of Actuaries, 8:182-192, 1920. [8] K. Nigam, A. K. Mccallum, S. Thrun, and T. Mitchell. Text classification fro'in labeled and unlabeled documents using EM. Machine Learning, 39:103-134, 2000. [9] W. Perks. Some observations on inverse probability, including a new indifference rule. Journal of the Institute of Actuaries, 73:285-312, 1947. [10] E. S. Ristad. A natural law ?of succession. Technical Report CS-TR-895-95, Department of Computer Science, Princeton University, 1995.
2063 |@word faculty:1 compression:9 proportion:3 essay:1 accounting:1 simplifying:1 tr:2 cleary:1 contains:1 document:1 outperforms:1 ka:2 com:1 informative:1 christian:3 remove:1 mccallum:1 sys:3 ith:1 gure:1 provides:3 along:1 direct:1 become:1 qualitative:1 consists:5 expected:3 rapid:1 behavior:2 abscissa:1 ol:1 terminal:1 informational:1 little:1 pf:6 cardinality:2 considering:2 window:3 becomes:1 estimating:5 panel:1 mass:1 kind:3 psych:1 string:1 differing:1 guarantee:1 quantitative:1 medical:1 producing:1 before:1 book2:1 despite:1 usenet:1 id:1 ap:1 might:2 examined:2 specifying:4 limited:4 range:2 testing:5 polity:4 procedure:2 featured:1 empirical:4 bell:1 word:26 griffith:1 unlabeled:4 prentice:1 context:3 applying:1 py:2 center:1 exposure:1 attention:1 wit:1 rule:3 notion:1 increment:1 laplace:6 target:1 us:3 distinguishing:1 hypothesis:3 harvard:1 element:1 rec:9 distributional:2 labeled:1 bottom:1 capture:1 calculate:2 news:2 removed:1 highest:2 valuable:1 disease:1 lidstone:2 intuition:1 nie:1 trained:2 ristad:7 compromise:1 upon:6 division:1 efficiency:1 baseball:2 completely:1 basis:2 translated:1 vid:2 represented:1 talk:13 distinct:2 effective:1 corresponded:1 oni:1 outside:2 header:1 stanford:3 otherwise:5 compressed:1 obj2:1 statistic:1 unseen:3 syntactic:1 itself:1 mb:1 motorcycle:3 flexibility:1 achieve:1 paper2:1 assessing:1 produce:1 object:2 illustrate:3 informs:1 ac:1 measured:1 transcript:1 soc:3 c:1 involves:1 paper5:1 bib:4 dio:1 ja:1 paper1:1 probable:1 pl:6 considered:1 hall:1 predict:1 dictionary:3 forsale:2 smallest:1 ilv:4 purpose:3 estimation:24 outperformed:2 label:1 always:1 rather:1 progc:1 encode:1 kid:1 improvement:1 consistently:1 likelihood:2 indicates:1 greatly:1 posteriori:1 inference:1 typically:2 classification:5 smoothing:1 special:1 equal:2 construct:1 once:2 cancel:1 others:1 report:2 duplicate:1 employ:1 consisting:3 friedman:10 attempt:1 progp:1 interest:1 paper3:1 weakness:1 punctuation:1 pc:1 behind:1 accurate:2 fu:1 necessary:1 solicited:2 addisonwesley:1 logarithm:1 mk:1 instance:1 classify:1 book1:2 newgroups:1 modeling:1 geo:2 mac:2 subset:3 uniform:1 conducted:1 graphic:2 st:2 ie:1 together:1 ivj:1 containing:4 actuary:2 worse:1 book:1 expert:6 potential:1 summarized:1 coding:1 includes:1 explicitly:2 caused:1 depends:1 performed:2 bayes:3 hf:1 contribution:1 formed:1 ni:5 who:1 succession:2 judgment:1 sy:1 weak:5 bayesian:12 famous:1 identification:1 modelled:1 produced:3 comp:8 solicit:1 published:1 classified:2 ty:8 frequency:3 naturally:1 attributed:1 paper6:1 dataset:3 mise:2 mitchell:1 knowledge:21 actually:1 higher:1 originally:1 follow:1 reflected:1 specify:1 improved:1 box:2 symptom:1 just:3 governing:1 hand:2 ei:1 ild:5 o:2 quality:1 indicated:1 facilitate:3 contain:1 consisted:1 true:2 ranged:1 inductive:1 assigned:2 jbt:1 misc:7 eg:1 oc:2 m:2 tt:1 demonstrate:1 oid:2 performs:1 fj:1 ef:1 novel:5 common:1 witten:1 multinomial:21 nh:2 occurred:1 ai:3 outlined:3 language:8 had:3 posterior:4 recent:1 belongs:1 verlag:1 formatted:2 binary:1 joshua:1 seen:1 unrestricted:1 preceding:1 eo:1 ey:1 employed:1 ii:1 full:1 desirable:2 technical:2 cross:3 long:2 divided:1 concerning:1 qi:1 prediction:5 involving:1 ko:11 normalization:1 goodman:1 ivi:2 file:22 nv:1 capitalization:1 tend:2 db:1 obj:1 structural:2 enough:1 bid:1 newsgroups:15 psychology:1 gave:2 restrict:1 identified:1 imperfect:1 idea:1 politics:7 defaulting:1 utility:1 dramatically:1 useful:1 paper4:1 listed:1 tenenbaum:1 ten:7 hardware:4 outperform:1 exist:1 dotted:1 estimated:7 v1a:1 diagnosis:1 discrete:2 write:1 promise:1 group:1 drawn:2 preprocessed:1 pj:5 sum:3 run:2 facilitated:1 inverse:1 uncertainty:7 prog:1 perk:5 bit:1 followed:1 replaces:1 encountered:2 occur:2 constraint:1 progl:1 gruffydd:1 department:2 designated:1 according:1 poor:1 conjugate:1 across:1 em:1 character:13 lid:2 making:2 jeffreys:5 restricted:6 pr:5 invariant:1 taken:1 resource:1 equation:4 remains:1 singer:10 tractable:1 end:1 hierarchical:1 appropriate:2 thomas:1 top:1 dirichlet:4 include:1 running:1 graphical:2 log2:2 maintaining:2 calculating:1 exploit:1 build:1 society:1 occurs:1 traditional:1 div:3 mapped:2 geological:1 thrun:1 gun:2 extent:1 length:4 code:1 relationship:1 illustration:1 providing:1 stated:1 unknown:2 perform:2 allowing:1 contributed:2 observation:4 extended:1 arbitrary:1 obj1:1 calgary:3 pic:2 namely:1 required:1 specified:2 philosophical:1 trans:2 suggested:2 bar:1 below:1 sparsity:1 program:1 built:1 including:3 royal:1 natural:6 predicting:2 representing:1 improve:2 technology:1 picture:1 lk:1 axis:1 fro:1 auto:3 text:17 prior:17 byte:4 determining:1 relative:1 law:9 sufficient:1 consistent:1 article:8 ascii:1 share:1 classifying:1 ibm:2 pi:1 token:1 free:1 english:4 side:2 allow:3 institute:1 taking:1 sparse:9 default:3 vocabulary:63 xn:6 world:1 stand:1 gram:1 author:1 commonly:1 adaptive:1 simplified:1 transaction:1 excess:1 approximate:3 corpus:4 table:2 hockey:1 ca:1 nigam:1 excellent:1 complex:1 suc:1 domain:13 arise:1 hyperparameters:1 atheism:3 allowed:1 join:1 ny:1 pv:3 wish:1 xl:1 candidate:3 lie:1 ib:1 printing:1 mideast:6 formula:1 specific:3 symbol:12 offset:1 alt:3 effectively:1 importance:1 illustrates:2 occurring:1 chen:1 easier:1 entropy:5 likely:1 expressed:1 contained:2 religion:6 indifference:1 sport:3 springer:1 determines:1 presentation:1 consequently:2 aided:1 included:2 specifically:1 determined:1 total:1 la:1 newsgroup:6 tiao:1 latter:1 assessed:1 evaluate:1 princeton:1 tested:1
1,167
2,064
Learning from Infinite Data in Finite Time Pedro Domingos Geoff H ulten Department of Computer Science and Engineering University of Washington Seattle, WA 98185-2350, U.S.A. {pedrod, ghulten} @cs.washington.edu Abstract We propose the following general method for scaling learning algorithms to arbitrarily large data sets. Consider the model Mii learned by the algorithm using ni examples in step i (ii = (nl , ... ,nm )) , and the model Moo that would be learned using infinite examples. Upper-bound the loss L(Mii' M oo ) between them as a function of ii, and then minimize the algorithm's time complexity f(ii) subject to the constraint that L(Moo , M ii ) be at most f with probability at most 8. We apply this method to the EM algorithm for mixtures of Gaussians. Preliminary experiments on a series of large data sets provide evidence of the potential of this approach. 1 An Approach to Large-Scale Learning Large data sets make it possible to reliably learn complex models. On the other hand , they require large computational resources to learn from. While in the past the factor limit ing the quality of learnable models was typically the quantity of data available, in many domains today data is super-abundant, and the bottleneck is t he t ime required to process it. Many algorithms for learning on large data sets have been proposed, but in order to achieve scalability they generally compromise the quality of the results to an unspecified degree. We believe this unsatisfactory state of affairs is avoidable, and in this paper we propose a general method for scaling learning algorithms to arbitrarily large databases without compromising the quality of the results. Our method makes it possible to learn in finite time a model that is essentially indistinguishable from the one that would be obtained using infinite data. Consider the simplest possible learning problem: estimating the mean of a random variable x. If we have a very large number of samples, most of them are probably superfluous. If we are willing to accept an error of at most f with probability at most 8, Hoeffding bounds [4] (for example) tell us that, irrespective of the distribution of x, only n = ~(R/f)2 1n (2/8) samples are needed, where R is x's range. We propose to extend this type of reasoning beyond learning single parameters, to learning complex models. The approach we propose consists of three steps: 1. Derive an upper bound on the relative loss between the finite-data and infinite-data models, as a function of the number of samples used in each step of the finite-data algorithm. 2. Derive an upper bound on the time complexity of the learning algorithm , as a function of the number of samples used in each step. 3. Minimize the time bound (via the number of samples used in each step) subject to target limits on the loss. In this paper we exemplify this approach using the EM algorithm for mixtures of Gaussians. In earlier papers we applied it (or an earlier version of it) to decision tree induction [2J and k-means clustering [3J. Despite its wide use, EM has long been criticized for its inefficiency (see discussion following Dempster et al. [1]), and has been considered unsuitable for large data sets [8J. Many approaches to speeding it up have been proposed (see Thiesson et al. [6J for a survey) . Our method can be seen as an extension of progressive sampling approaches like Meek et al. [5J: rather than minimize the total number of samples needed by the algorithm , we minimize the number needed by each step, leading to potentially much greater savings; and we obtain guarantees that do not depend on unverifiable extrapolations of learning curves. 2 A Loss Bound for EM In a mixture of Gaussians model, each D-dimensional data point Xj is assumed to have been independently generated by the following process: 1) randomly choose a mixture component k; 2) randomly generate a point from it according to a Gaussian distribution with mean f-Lk and covariance matrix ~k. In this paper we will restrict ourselves to the case where the number K of mixture components and the probability of selection P(f-Lk) and covariance matrix for each component are known. Given a training set S = {Xl, ... , XN }, the learning goal is then to find the maximumlikelihood estimates of the means f-Lk. The EM algorithm [IJ accomplishes this by, starting from some set of initial means , alternating until convergence between estimating the probability p(f-Lk IXj) that each point was generated by each Gaussian (the Estep), and computing the ML estimates of the means ilk = 2::;':1 WjkXj / 2::f=l Wjk (the M step), where Wjk = p(f-Lklxj) from the previous E step. In the basic EM algorithm, all N examples in the training set are used in each iteration. The goal in this paper is to speed up EM by using only ni < N examples in the ith iteration, while guaranteeing that the means produced by the algorithm do not differ significantly from those that would be obtained with arbitrarily large N. Let Mii = (ill , . . . , ilK) be the vector of mean estimates obtained by the finite-data EM algorithm (i.e., using ni examples in iteration i), and let Moo = (f-L1, ... ,f-LK) be the vector obtained using infinite examples at each iteration. In order to proceed, we need to quantify the difference between Mii and Moo . A natural choice is the sum of the squared errors between corresponding means, which is proportional to the negative log-likelihood of the finite-data means given the infinite-data ones: K L(Mii' Moo ) = L k=l K Ililk - f-Lkl12 = D LL lilkd - f-Lkdl 2 (1) k=ld=l where ilkd is the dth coordinate of il, and similarly for f-Lkd. After any given iteration of EM, lilkd - f-Lkdl has two components. One, which we call the sampling error, derives from the fact that ilkd is estimated from a finite sample, while J-Lkd is estimated from an infinite one. The other component, which we call the weighting error, derives from the fact that , due to sampling errors in previous iterations, the weights Wjk used to compute the two estimates may differ. Let J-Lkdi be the infinite-data estimate of the dth coordinate of the kth mean produced in iteration i, flkdi be the corresponding finite-data estimate, and flkdi be the estimate that would be obtained if there were no weighting errors in that iteration. Then the sampling error at iteration i is Iflkdi - J-Lkdi I, the weighting error is Iflkdi - flkdi I, and the total error is Iflkdi - J-Lkdi 1 :::; Iflkdi - flkdi 1 + Iflkdi - J-Lkdi I? Given bounds on the total error of each coordinate of each mean after iteration i-I, we can derive a bound on the weighting error after iteration i as follows. Bounds on J-Lkd ,i-l for each d imply bounds on p(XjlJ-Lki ) for each example Xj, obtained by substituting the maximum and minimum allowed distances between Xjd and J-Lkd ,i-l into the expression of the Gaussian distribution. Let P}ki be the upper bound on P(XjlJ-Lki) , and Pjki be the lower bound. Then the weight of example Xj in mean J-Lki can be bounded from below by by W}ki W (+) -jki = min{p}kiP(J-Lk)/ wjki = PjkiP(J-Lk)/ ~~= l P}k'iP(J-LU, and from above ~~=l Pjk'iP(J-LU, I}. Let w;t: = W}ki if Xj ::::: 0 and (- ) -- W jki -'f > 0 an d W jki (- ) -- W jki + 0 th erWlse. . W jki 1 Xj _ ' ot h erWlse, an d 1et W- jki Then Iflkdi - flkdi , IJ-Lkdi 1 < max - ~7~1 Wjk i Xj I ",ni uj=l {I , J-Lkdi - Wjki uj =l " , ni W jki (+) Xj ",ni _ uj=l w jki II , ,J-Lkdi - uj =l ",ni ( - ) Xj W jki ",ni uj=l + I} (2) w jki A corollary of Hoeffding's [4] Theorem 2 is that, with probability at least 1 - 8, the sampling error is bounded by Iflkdi - J-Lkdi 1 :::; (3) where Rd is the range of the dth coordinate of the data (assumed known 1 ). This bound is independent of the distribution of the data, which will ensure that our results are valid even if the data was not truly generated by a mixture of Gaussians, as is often the case in practice. On the other hand, the bound is more conservative than distribution-dependent ones, requiring more samples to reach the same guarantees. The initialization step is error-free, assuming the finite- and infinite-data algorithms are initialized with the same means. Therefore the weighting error in the first iteration is zero, and Equation 3 bounds the total error. From this we can bound the weighting error in the second iteration according to Equation 2, and therefore bound the total error by the sum of Equations 2 and 3, and so on for each iteration until the algorithms converge. If the finite- and infinite-data EM converge in the same number of iterations m, the loss due to finite data is L(Mii" Moo ) = ~f= l ~~= llflkdm - J-Lkdml 2 (see Equation 1). Assume that the convergence criterion is ~f= l IIJ-Lki - J-Lk ,i- 111 2 :::; f. In general 1 Although a normally distributed variable has infinite range, our experiments show that assuming a sufficiently wide finite range does not significantly affect the results. (with probability specified below), infinite-data EM converges at one of the iterations for which the minimum possible change in mean positions is below ,,/, and is guaranteed to converge at the first iteration for which the maximum possible change is below "(. More precisely, it converges at one of the iterations for which ~~= l ~~= l (max{ IPkd ,i- l - Pkdil-IPkd,i - l - ftkd,i - ll-IPkdi - ftkdil, O})2 ::; ,,/, and is guaranteed to converge at the first iteration for which ~~=l ~~=l (IPkd,i-l Pkdil + IPkd ,i-l - ftkd ,i-ll + IPkdi - ftkdil)2 ::; "/. To obtain a bound for L(Mn, Moo), finite-data EM must be run until the latter condition holds. Let I be the set of iterations at which infinite-data EM could have converged. Then we finally obtain where m is the total number of iterations carried out. This bound holds if all of the Hoeffding bounds (Equation 3) hold. Since each of these bounds fails with probability at most 8, the bound above fails with probability at most 8* = K Dm8 (by the union bound). As a result, the growth with K, D and m of the number of examples required to reach a given loss bound with a given probability is only O(v'lnKDm). The bound we have just derived utilizes run-time information, namely the distance of each example to each mean along each coordinate in each iteration. This allows it to be tighter than a priori bounds. Notice also that it would be trivial to modify the treatment for any other loss criterion that depends only on the terms IPkdm - ftkdm I (e.g., absolute loss) . 3 A Fast EM Algorithm We now apply the previous section's result to reduce the number of examples used by EM at each iteration while keeping the loss bounded. We call the resulting algorithm VFEM. The goal is to learn in minimum time a model whose loss relative to EM applied to infinite data is at most f* with probability at least 1 - 8*. (The reason to use f* instead of f will become apparent below.) Using the notation of the previous section, if ni examples are used at each iteration then the running time of EM is O(KD ~::l ni) , and can be minimized by minimizing ~::l ni. Assume for the moment that the number of iterations m is known. Then, using Equation 1, we can state the goal more precisely as follows. Goal: Minimize ~::l ni, subject to the constraint that ~~=l IIPkm - ftkml12 ::; f* with probability at least 1 - 8* . A sufficient condition for ~~=l IIPkm - ftkml12 ::; f* is that Vk IIPkm - ftkmll ::; Jf*/K. We thus proceed by first minimizing ~::l ni subject to IIPkm - ftkmll ::; f* / K separately for each mean. 2 In order to do this, we need to express IIPkm ftkm II as a function of the ni 'so By the triangle inequality, IIPki - ftki II ::; IIPki - ftki II + J Ilftki - ftk& By Equation 3, Ilftki - ftki II::; ~R2ln(2/8) ~;~ l w;kd(~;~ l Wjki)2, where R2 = ~~=l RJ and 8 = 8* / K Dm per the discussion following Equation 4. The (~;~ l Wjki)2 / ~;~ l W;ki term is a measure of the diversity of the weights , 2This will generally lead to a suboptimal solution; improving it is a matter for future work. being equal to 1 - l/Gini(W~i)' where W~i is the vector of normalized weights wjki = wjkd 2:j,i=l Wjl ki. It attains a minimum of! when all the weights but one are zero, and a maximum of ni when all the weights are equal and non-zero. However, we would like to have a measure whose maximum is independent of ni, so that it remains approximately constant whatever the value of ni chosen (for sufficiently large ni). The measure will then depend only on the underlying distribution of the data. Thus we define f3ki = (2:7~1 Wjki)2 /(ni 2:7~1 W]ki) ' obtaining IliLki - ILkill :::; JR 2 ln (2/8)/(2f3ki n i). Also, IIP-ki-iLkill = J2:~= llP-kdi - iLkdil 2, with lP-kdi-iLkdil bounded by Equation 2. To keep the analysis tractable, we upper-bound this term by a function proportional to IIP-kd,i-1 - ILkd,i-111. This captures the notion than the weighting error in one iteration should increase with the total error in the previous one. Combining this with the bound for IliLki - ILkill, we obtain R 2 l n (2/8) 2f3kini (5) where CXki is the proportionality constant. Given this equation and IIP-kO - ILkO II = 0, it can be shown by induction that m IIP-km - ILkmll :::; ~~ (6) where (7) The target bound will thus be satisfied by minimizing 2:: 1 ni subject to 2::1 (rkd,;niJ = J E* / K. 3 Finding the n/s by the method of Lagrange multipliers yields ni = ~ (f ~rkir%j) 2 (8) )=1 This equation will produce a required value of ni for each mean. To guarantee the desired E*, it is sufficient to make ni equal to the maximum of these values. The VFEM algorithm consists of a sequence of runs of EM, with each run using more examples than the last, until the bound L(Mii' Moo) :::; E* is satisfied, with L(Mii' Moo) bounded according to Equation 4. In the first run, VFEM postulates a maximum number of iterations m, and uses it to set 8 = 8* / K Dm. If m is exceeded, for the next run it is set to 50% more than the number needed in the current run. (A new run will be carried out if either the 8* or E* target is not met.) The number of examples used in the first run of EM is the same for all iterations, and is set to 1.1(K/2)(R/E*)2ln(2/8). This is 10% more than the number of examples that would theoretically be required in the best possible case (no weighting errors in the last 3This may lead to a suboptimal solution for the ni's, in the unlikely case that II increases with them. Jtkm Ilflkm - iteration, leading to a pure Hoeffding bound, and a uniform distribution of examples among mixture components). The numbers of examples for subsequent runs are set according to Equation 8. For iterations beyond the last one in the previous run , the number of examples is set as for the first run. A run of EM is terminated when L~= l L~= l (Iflkd ,i- l - flkdi 1+ Iflkd ,i-l - ILkd ,i-l l + Iflkdi - ILkdi 1)2 :s: "( (see discussion preceding Equation 4), or two iterations after L~=l IIILki - ILk,i-1 11 2 :s: "( 13, whichever comes first. The latter condition avoids overly long unproductive runs. If the user target bound is E, E* is set to min{ E, "( 13}, to facilitate meeting the first criterion above. When the convergence threshold for infinite-data EM was not reached even when using the whole training set, VFEM reports that it was unable to find a bound; otherwise the bound obtained is reported. VFEM ensures that the total number of examples used in one run is always at least twice the number n used in the previous run. This is done by, if L ni < 2n, setting the ni's instead to n~ = 2n(nil L ni). If at any point L ni > mN, where m is the number of iterations carried out and N is the size of the full training set, Vi ni = N is used. Thus, assuming that the number of iterations does not decrease with the number of examples, VFEM's total running time is always less than three times the time taken by the last run of EM. (The worst case occurs when the one-but-last run is carried out on almost the full training set.) The run-time information gathered in one run is used to set the n/s for the next run. We compute each Ctki as Ilflki - Pkill/llflk ,i-l - ILk ,i-lll. The approximations made in the derivation will be good, and the resulting ni's accurate, if the means' paths in the current run are similar to those in the previous run. This may not be true in the earlier runs , but their running time will be negligible compared to that of later runs , where the assumption of path similarity from one run to the next should hold. 4 Experiments We conducted a series of experiments on large synthetic data sets to compare VFEM with EM. All data sets were generated by mixtures of spherical Gaussians with means ILk in the unit hypercube. Each data set was generated according to three parameters: the dimensionality D , the number of mixture components K , and the standard deviation (Y of each coordinate in each component. The means were generated one at a time by sampling each dimension uniformly from the range (2(Y,1 - 2(Y). This ensured that most of the data points generated were within the unit hypercube. The range of each dimension in VFEM was set to one. Rather than discard points outside the unit hypercube, we left them in to test VFEM's robustness to outliers. Any ILk that was less than (vD1K)(Y away from a previously generated mean was rejected and regenerated, since problems with very close means are unlikely to be solvable by either EM or VFEM. Examples were generated by choosing one of the means ILk with uniform probability, and setting the value of each dimension of the example by randomly sampling from a Gaussian distribution with mean ILkd and standard deviation (Y. We compared VFEM to EM on 64 data sets of 10 million examples each, generated by using every possible combination of the following parameters: D E {4, 8,12, 16}; K E {3, 4, 5, 6} ; (Y E {.01 , .03, .05, .07}. In each run the two algorithms were initialized with the same means, randomly selected with the constraint that no two be less than vD1(2K) apart. VFEM was allowed to converge before EM's guaranteed convergence criterion was met (see discussion preceding Equation 4). All experiments were run on a 1 GHz Pentium III machine under Linux, with "( = O.OOOlDK, 8* = 0.05, and E* = min{O.Ol, "(} . Table 1: Experimental results. Values are averages over the number of runs shown. Times are in seconds, and #EA is the total number of example accesses made by the algorithm, in millions. Runs Bound No bound All Algorithm VFEM EM VFEM EM VFEM EM #Runs 40 40 24 24 64 64 Time 217 3457 7820 4502 3068 3849 #EA 1.21 19.75 43.19 27.91 16.95 22.81 Loss 2.51 2.51 1.20 1.20 2.02 2.02 D 10.5 10.5 9.1 9.1 10 10 K 4.2 4.2 4.9 4.9 4.5 4.5 rr 0.029 0.029 0.058 0.058 0.04 0.04 The results are shown in Table 1. Losses were computed relative to the true means, with the best match between true means and empirical ones found by greedy search. Results for runs in which VFEM achieved and did not achieve the required E* and 8* bounds are reported separately. VFEM achieved the required bounds and was able to stop early on 62.5% of its runs. When it found a bound, it was on average 16 t imes faster than EM. When it did not, it was on average 73% slower. The losses of the two algorithms were virtually identical in both situations. VFEM was more likely to converge rapidly for higher D's and lower K's and rr's. When achieved , the average loss bound for VFEM was 0.006554, and for EM it was 0.000081. In other words, the means produced by both algorithms were virtually identical to those that would be obtained with infinite data. 4 We also compared VFEM and EM on a large real-world data set, obtained by recording a week of Web page requests from the entire University of Washington campus. The data is described in detail in Wolman et al. [7], and the preprocessing carried out for these experiments is described in Domingos & Hulten [3]. The goal was to cluster patterns of Web access in order to support distributed caching. On a dataset with D = 10 and 20 million examples, with 8* = 0.05, I = 0.001, E* = 1/3, K = 3, and rr = 0.01, VFEM achieved a loss bound of 0.00581 and was two orders of magnitude faster than EM (62 seconds vs. 5928), while learning essentially the same means. VFEM's speedup relative to EM will generally approach infinity as the data set size approaches infinity. The key question is thus: what are the data set sizes at which VFEM becomes worthwhile? The tentative evidence from these experiments is that they will be in the millions. Databases of this size are now common, and their growth continues unabated , auguring well for the use of VFEM. 5 Conclusion Learning algorithms can be sped up by minimizing the number of examples used in each step, under the constraint that the loss between the resulting model and the one that would be obtained with infinite data remain bounded. In this paper we applied this method to the EM algorithm for mixtures of Gaussians, and observed the resulting speedups on a series of large data sets. 4The much higher loss values relative to the true means, however, indicate that infinitedata EM would often find only local optima (unless the greedy search itself only found a suboptimal match). Acknowledgments This research was partly supported by NSF CAREER and IBM Faculty awards to the first author, and by a gift from the Ford Motor Company. References [1] A. P. Dempster, N. M. Laird, and D. B. Rubin. Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society, Series B, 39:1- 38, 1977. [2] P. Domingos and G. Hulten. Mining high-speed data streams. In Proceedings of the Sixth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 71- 80, Boston, MA, 2000. ACM Press. [3] P. Domingos and G. Hulten. A general method for scaling up machine learning algorithms and its application to clustering. In Proceedings of the Eighteenth International Conference on Machine Learning, pp. 106-113, Williamstown, MA, 2001. Morgan Kaufmann. [4] W. Hoeffding. Probability inequalities for sums of bounded random variables. Journal of the American Statistical Association, 58:13- 30, 1963. [5] C. Meek, B. Thiesson, and D. Heckerman. The learning-curve method applied to clustering. Technical Report MSR-TR-01-34, Microsoft Research, Redmond, WA,2000. [6] B. Thiesson, C. Meek, and D. Heckerman. Accelerating EM for large databases. Technical Report MSR-TR-99-31, Microsoft Research, Redmond, WA, 2001. [7] A. Wolman, G. Voelker, N. Sharma, N. Cardwell, M. Brown, T. Landray, D. Pinnel, A. Karlin, and H. Levy. Organization-based analysis of Web-object sharing and caching. In Proceedings of the Second USENIX Conference on Internet Technologies and Systems, pp. 25- 36, Boulder, CO, 1999. [8] T. Zhang, R. Ramakrishnan, and M. Livny. BIRCH: An efficient data clustering method for very large databases. In Proceedings of the 1996 A CM SIGMOD International Conference on Management of Data, pp. 103- 114, Montreal, Canada, 1996. ACM Press.
2064 |@word msr:2 faculty:1 version:1 proportionality:1 willing:1 km:1 covariance:2 tr:2 ld:1 moment:1 initial:1 inefficiency:1 series:4 past:1 current:2 ixj:1 moo:9 must:1 subsequent:1 motor:1 v:1 greedy:2 selected:1 affair:1 ith:1 zhang:1 along:1 become:1 consists:2 theoretically:1 ol:1 spherical:1 company:1 lll:1 becomes:1 gift:1 estimating:2 bounded:7 notation:1 underlying:1 campus:1 what:1 cm:1 unspecified:1 livny:1 finding:1 guarantee:3 every:1 growth:2 ensured:1 whatever:1 normally:1 unit:3 before:1 negligible:1 engineering:1 local:1 modify:1 limit:2 despite:1 path:2 approximately:1 twice:1 initialization:1 jki:10 co:1 range:6 wjl:1 acknowledgment:1 practice:1 union:1 empirical:1 significantly:2 word:1 close:1 selection:1 eighteenth:1 starting:1 independently:1 survey:1 pure:1 notion:1 coordinate:6 target:4 today:1 user:1 us:1 domingo:4 continues:1 database:4 observed:1 capture:1 worst:1 imes:1 ensures:1 ftki:3 decrease:1 dempster:2 complexity:2 depend:2 compromise:1 triangle:1 geoff:1 derivation:1 fast:1 gini:1 tell:1 outside:1 choosing:1 whose:2 apparent:1 voelker:1 otherwise:1 itself:1 ford:1 ip:2 laird:1 sequence:1 rr:3 karlin:1 propose:4 j2:1 combining:1 rapidly:1 achieve:2 scalability:1 wjk:4 seattle:1 convergence:4 cluster:1 optimum:1 produce:1 guaranteeing:1 converges:2 object:1 oo:1 derive:3 montreal:1 ij:2 c:1 come:1 indicate:1 quantify:1 differ:2 met:2 compromising:1 require:1 pjk:1 preliminary:1 tighter:1 extension:1 hold:4 sufficiently:2 considered:1 week:1 substituting:1 early:1 gaussian:4 always:2 super:1 rather:2 caching:2 hulten:3 corollary:1 derived:1 vk:1 unsatisfactory:1 likelihood:2 pentium:1 sigkdd:1 attains:1 dependent:1 typically:1 unlikely:2 accept:1 entire:1 among:1 ill:1 priori:1 xjd:1 equal:3 saving:1 washington:3 sampling:7 identical:2 progressive:1 future:1 minimized:1 report:3 regenerated:1 randomly:4 ime:1 ourselves:1 microsoft:2 organization:1 mining:2 mixture:10 truly:1 nl:1 superfluous:1 accurate:1 unless:1 tree:1 incomplete:1 initialized:2 abundant:1 desired:1 nij:1 criticized:1 earlier:3 deviation:2 uniform:2 conducted:1 reported:2 synthetic:1 international:3 linux:1 squared:1 postulate:1 nm:1 satisfied:2 iip:4 choose:1 management:1 hoeffding:5 american:1 leading:2 potential:1 diversity:1 matter:1 depends:1 vi:1 stream:1 later:1 extrapolation:1 lkd:4 reached:1 minimize:5 lkl12:1 ni:30 il:1 kaufmann:1 yield:1 gathered:1 produced:3 lu:2 converged:1 reach:2 sharing:1 sixth:1 pp:4 dm:2 stop:1 dataset:1 treatment:1 birch:1 exemplify:1 knowledge:1 dimensionality:1 ea:2 exceeded:1 higher:2 done:1 just:1 rejected:1 until:4 hand:2 web:3 quality:3 believe:1 facilitate:1 brown:1 multiplier:1 true:4 requiring:1 normalized:1 alternating:1 indistinguishable:1 ll:3 criterion:4 l1:1 reasoning:1 common:1 thiesson:3 sped:1 million:4 extend:1 he:1 association:1 rd:1 similarly:1 access:2 similarity:1 rkd:1 discard:1 apart:1 inequality:2 arbitrarily:3 meeting:1 morgan:1 seen:1 minimum:4 greater:1 preceding:2 accomplishes:1 converge:6 sharma:1 ii:11 full:2 rj:1 ing:1 technical:2 match:2 faster:2 long:2 award:1 basic:1 ko:1 essentially:2 iteration:33 achieved:4 separately:2 ot:1 probably:1 subject:5 recording:1 virtually:2 call:3 iii:1 xj:8 affect:1 restrict:1 suboptimal:3 reduce:1 bottleneck:1 expression:1 accelerating:1 vd1:1 proceed:2 generally:3 simplest:1 generate:1 nsf:1 notice:1 estimated:2 overly:1 per:1 express:1 key:1 threshold:1 sum:3 run:33 almost:1 utilizes:1 mii:8 decision:1 scaling:3 bound:41 ki:7 internet:1 meek:3 guaranteed:3 ilk:7 constraint:4 precisely:2 infinity:2 ftk:1 unverifiable:1 speed:2 min:3 estep:1 department:1 speedup:2 according:5 combination:1 request:1 kd:3 jr:1 remain:1 heckerman:2 em:38 lp:1 outlier:1 boulder:1 taken:1 ln:2 resource:1 equation:15 remains:1 previously:1 kdi:2 needed:4 tractable:1 whichever:1 available:1 gaussians:6 apply:2 worthwhile:1 away:1 robustness:1 slower:1 clustering:4 ensure:1 running:3 unsuitable:1 sigmod:1 uj:5 hypercube:3 society:1 question:1 quantity:1 occurs:1 kth:1 ilkd:5 distance:2 unable:1 trivial:1 reason:1 induction:2 assuming:3 minimizing:4 potentially:1 lki:4 negative:1 reliably:1 upper:5 finite:13 situation:1 usenix:1 canada:1 namely:1 required:6 specified:1 kip:1 tentative:1 learned:2 beyond:2 dth:3 llp:1 below:5 able:1 pattern:1 redmond:2 max:2 royal:1 natural:1 solvable:1 mn:2 pedrod:1 technology:1 imply:1 lk:8 irrespective:1 carried:5 speeding:1 discovery:1 relative:5 loss:17 proportional:2 degree:1 sufficient:2 rubin:1 ibm:1 supported:1 last:5 free:1 keeping:1 wide:2 absolute:1 distributed:2 ghz:1 curve:2 dimension:3 xn:1 valid:1 avoids:1 world:1 author:1 made:2 preprocessing:1 wolman:2 keep:1 ml:1 assumed:2 search:2 table:2 learn:4 erwlse:2 career:1 obtaining:1 improving:1 complex:2 domain:1 did:2 terminated:1 whole:1 allowed:2 iij:1 fails:2 position:1 xl:1 levy:1 weighting:8 theorem:1 learnable:1 r2:1 evidence:2 derives:2 magnitude:1 boston:1 likely:1 lagrange:1 pedro:1 williamstown:1 ramakrishnan:1 avoidable:1 acm:3 ma:2 goal:6 jf:1 change:2 infinite:17 uniformly:1 conservative:1 total:10 nil:1 partly:1 experimental:1 support:1 maximumlikelihood:1 latter:2
1,168
2,065
Probabilistic Abstraction Hierarchies Eran Segal Computer Science Dept. Stanford University eran@cs.stanford.edu Daphne Koller Computer Science Dept. Stanford University koller@cs.stanford.edu Dirk Ormoneit Computer Science Dept. Stanford University ormoneit@cs.stanford.edu Abstract Many domains are naturally organized in an abstraction hierarchy or taxonomy, where the instances in ?nearby? classes in the taxonomy are similar. In this paper, we provide a general probabilistic framework for clustering data into a set of classes organized as a taxonomy, where each class is associated with a probabilistic model from which the data was generated. The clustering algorithm simultaneously optimizes three things: the assignment of data instances to clusters, the models associated with the clusters, and the structure of the abstraction hierarchy. A unique feature of our approach is that it utilizes global optimization algorithms for both of the last two steps, reducing the sensitivity to noise and the propensity to local maxima that are characteristic of algorithms such as hierarchical agglomerative clustering that only take local steps. We provide a theoretical analysis for our algorithm, showing that it converges to a local maximum of the joint likelihood of model and data. We present experimental results on synthetic data, and on real data in the domains of gene expression and text. 1 Introduction Many domains are naturally associated with a hierarchical taxonomy, in the form of a tree, where instances that are close to each other in the tree are assumed to be more ?similar? than instances that are further away. In biological systems, for example, creating a taxonomy of the instances is often one of the first steps in understanding the system. In particular, much of the work on analyzing gene expression data [3] has focused on creating gene hierarchies. Similarly, in text domains, creating a hierarchy of documents is a common task [12, 7]. In many of these applications, the hierarchy is unknown; indeed, discovering the hierarchy is often a key part of the analysis. The standard algorithms applied to the problem typically use an agglomerative bottom-up approach [3] or a divide-and-conquer top-down approach [8]. Although these methods have been shown to be useful in practice, they suffer from several limitations: First, they proceed via a series of local improvements, making them particularly prone to local maxima. Second, at least the bottom-up approaches are typically applied to the raw data; models (if any), are constructed as a post-processing step. Thus, domain knowledge about the type of distribution from which data instances are sampled is rarely used in the formation of the hierarchy. In this paper, we present probabilistic abstraction hierarchies (PAH), a probabilistically principled general framework for learning abstraction hierarchies from data which overcomes these difficulties. We use a Bayesian approach, where the different models correspond to different abstraction hierarchies. The prior is designed to enforce our intuitions about taxonomies: nearby classes have similar data distributions. More specifically, a model in a PAH is a tree, where each node in the tree is associated with a class-specific probabilistic model (CPM). Data is generated only at the leaves of the tree, so that a model basically defines a mixture distribution whose components are the CPMs at the leaves of the tree. The CPMs at the internal nodes are used to define the prior over models: We prefer models where the CPM at a child node is close to the CPM at its parent, relative to some distance function between CPMs. Our framework allows a wide range of notions of distance between models; we essentially require only that the distance function be convex in the parameters of the two CPMs. For example, if a CPM is a Gaussian distribution, we might use a simple squared Euclidean distance between the parameters of the two CPMs. We present a novel algorithm for learning the model parameters and the tree structure in this framework. Our algorithm is based on the structural EM (SEM) approach of [4], but utilizes ?global? optimization steps for learning the best possible hierarchy and CPM parameters (see also [5, 13] for similar global optimization steps within SEM). Each step in our procedure is guaranteed to increase the joint probability of model and data, and hence (like SEM) our procedure is guaranteed to converge to a local optimum. Our approach has several advantages. (1) It provides principled probabilistic semantics for hierarchical models. (2) It is model based, which allows us to exploit domain structural knowledge more easily. (3) It utilizes global optimization steps, which tend to avoid local maxima and help make the model less sensitive to noise. (4) The abstraction hierarchy tends to pull the parameters of one model closer to those of nearby ones, which naturally leads to a form of parameter smoothing or shrinkage [12]. We present experiments for PAH on synthetic data and on two real data sets: gene expression and text. Our results show that the PAH approach produces hierarchies that are more robust to noise in the data, and that the learned hierarchies generalize better to test data than those produced by hierarchical agglomerative clustering. 2 Probabilistic Abstraction Hierarchy Let be the domain of some random observation, e.g., the set of possible assignments to a set of features. Our goal is to take a set of instances in , and to cluster them into some set of classes. Standard ?flat? clustering approaches ? for example, Autoclass [1] or the -means algorithm ? are special cases of a generative mixture model. In such models, each data instance belongs to one of the classes, each of which is associated with a different class-specific probabilistic model (CPM). Each data instance is sampled independently by first selecting one of the classes according to a multinomial distribution, and then randomly selecting the data instance itself from the CPM of the chosen class. In standard clustering models, there is no relation between the individual CPMs, which can be arbitrarily different. In this paper, we propose a model where the different classes are related to each other via an abstraction hierarchy, such that classes that are ?nearby? in the hierarchy have similar probabilistic models. More precisely, we define:                    "    ! # # %   #   $    Our framework does not, in principle, place restrictions on the form of the CPMs; we can use any probabilistic model that defines a probability distribution over . For example, # may be a Bayesian network, in which case its specification would include  may bethea parameters, and perhaps also the network structure; in a different setting, # hidden Markov model. In practice, however, the choice of CPMs has ramifications both for the Definition 2.1 A probabilistic abstraction hierarchy (PAH) is a tree with nodes and undirected edges , such that has exactly leaves . Each node , , is associated with a CPM , which defines a distribution over ; we use to denote . We also have a multinomial distribution over the leaves ; we use to denote the parameters of this distribution. overall hierarchical model and the algorithm. As discussed above, we assume that data is generated only from the leaves of the tree. Thus, we augment with an additional hidden class variable for each data item, which takes the values denoting the leaf that was chosen to generate this item. Given a PAH , an element , and a value for , we define , where is the multinomial distribution over the leaves and is the conditional density of the data item given the CPM at leaf . The induced   %+-718,.+-',9/':#  /;< #=1 <>1 +-'),9( &?@*A/;%1 * & & +-,.' *0/2123+-,.&45*6/ * M5 (=M4) f(g2) M5 g1 g3 g2 M2 f(g2) f(g2) M4 M1 M4 M3 M4 g1 g2 g3 g1 g4 g2 g3 M2 f(g2) M5 (=M4) M3 M2 M1 M2 g2 g3 g1 g2 g3 M6 (=M4) M4 M1 g4 M3 g4 f(g2) g1 M6 (=M3) g4 g4 M6 (=M3) M1 M3 (a) (b) Figure 1: (a) A PAH with 3 leaves over a 4-dimensional continuous state space, along with a visualization of the Gaussian distribution for the 3rd dimension. (b) Two different weight-preserving transformations for a tree with 4 leaves . #   # ' +-,9 ' * / 21 +-,.'-/ 21 * # #  distribution of given , from which the data are generated, is simply , where is summed out from . As we mentioned, the role of the internal nodes in the tree is to enforce an intuitive interpretation of the model as an abstraction hierarchy, by enforcing similarity between CPMs at nearby leaves. We achieve this goal by defining a prior distribution over aband straction hierarchies that penalizes the distance between neighboring CPMs using a distance function . Note that we do not require that be a distance in the mathematical sense; instead, we only require that it be symmetric (as we chose to use undirected trees), non-negative, and that iff .1 One obvious choice is to define IDKL IDKL , where IDKL is the KLdistance between the distributions that and define over . This distance measure has the advantage of being applicable to any pair of CPMs over the same space, even if their parameterization is different. Given a definition of , we define the prior over PAHs as , where represents the extent to which differences in distances are penalized (larger represents a larger penalty). 2 Given a set of data instances with domain , our goal is to find a PAH that maximizes or equivalently, . By maximizing this expression, we are trading off the fit of the mixture model over the leaves to the data , and the desire to generate a hierarchy in which nearby models are similar. Fig. 1(a) illustrates a typical PAH with Gaussian CPM distributions, where a CPM close to the leaves of the tree is more specialized and thus has fairly peaked distributions. Conversely, CPMs closer to the root of the tree, acting to bridge between their neighbors, are expected to have less peaked distributions and peak only around parts of the distribution which are common to an entire subtree.  7 ,.# #  1 7 ,.# #  1- ,.# #  1 #   ,9# +-,.21     ,  ! ,9#  #  11 " +-,. /#" 1$ !+-, 21+-," / 21  #  1  #  #  ,# .#  > # 1 9, #>#  1   +-, 21 )%*&+( +-,"  / 21 %'&( " 3 Learning the Models # "   #     - ,/. 0  1,2. 3 0   %  " Our goal in this section is to learn a PAH from a data set . This learning task is fairly complex, as many aspects are unknown: the structure of the tree , the CPMs at the nodes of , the parameters , and the assignment of the instances in to leaves of . Hence, the likelihood function has multiple local maxima, and no general method exists for finding the global maximum. In this section, we provide an efficient algorithm for finding a locally optimal . 1   ; =KJLB 4!5 687:9<; =5?>A@CBEDF; =5 >A@HGIB Two models are considered identical if . Care must be taken to ensure that is a proper probability distribution, but this will always be the case for the choice of we use in this paper. We also note that, if desired, we can modify this prior to incorporate a prior over the parameters of the ?s. 2 M @:N #   #  #  To simplify the algorithm, we assume that the structure of the CPMs is fixed. This reduces the choice of each to a pure numerical optimization problem. The general framework of our algorithm extends to cases where we also have to solve the model selection problem for each , but the computational issues are somewhat different. We first discuss the case of complete data, where for each data instance , we are given the leaf from which it was generated. For this case, we show how to learn the structure of the tree and the setting of the parameters and . This problem, of constructing a tree over a set of points that is not fixed, is very closely related to the Steiner tree problem [10], virtually all of whose variants are NP-hard. We propose a heuristic approach that decouples the joint optimization problem into two subproblems: optimizing the CPM parameters given the tree structure, and learning a tree structure given a set of CPMs. Somewhat surprisingly, we show that our careful choice of additive prior allows each of these subproblems to be tackled very effectively using global optimization techniques. We begin with the task of learning the CPMs. Thus, assume that we are given both the data instance to one of the structure of the tree and the assignment of each  leaves, denoted . It remains to find that minimize  . Substituting the definitions into  , we get that #   2 1  %'&( +-,.21 &     D  % . " 0 ; = $ ,/. " 0 ( " %   $   /> !B     ,/. 0 ( "   N  !" $# N :  %'&( +-, " / ; = />-@ N B &%'   % )( % * N M =@ N @ B 2 43 ,+ -/.01 - /5 (1) The first term, involving the multinomial parameters , separates from the rest, so that the optimization of  relative to reduces to straightforward maximum likelihood estimation. To optimize the CPM parameters, the key property turns out to be the convexity of the  function, which holds in a wide variety of choices of CPMs and ; in particular, it holds for the models used in our experiments. The convexity property allows us to find the global minimum of  using a simple iterative procedure. In each iteration, we optimize the 76 parameters of one of the ?s, fixing the parameters of the remaining CPMs ( ). This procedure is repeated for each of the ?s in a round robin fashion, until convergence. By the joint convexity of  , this iterative procedure is guaranteed to converge to the global inminimum of  . An examination of (1) shows that the optimization of each CPM volves only the data cases assigned to (if is a leaf) and the parameters of the CPMs that are neighbors of in the tree, thereby simplifying the computation substantially. We now turn our attention to the second subproblem, of learning the structure of the tree given the learned CPMs. We first consider an empty tree containing only the (unconnected) leaf nodes , and find the optimal parameter settings for each leaf CPM as described above. Note that these CPMs are unrelated, and the parameters of each one are computed independently of other CPMs. Given this initial set of CPMs for the leaf nodes , the algorithm tries to learn a good tree structure relative to these CPMs. The goal is to find the lowest weight tree, subject to the restriction that the tree structure must keep the same set of leaves . Due to the decomposability of , the . This probpenalty of the tree can be measured via the sum of the edge weights lem is also a variant of the Steiner tree problem. As a heuristic substitute, we follow the lines of [5] and use a minimum spanning tree (MST) algorithm for constructing low-weight trees. At each iteration, the algorithm starts out with a tree over some set of nodes . of this tree, and constructs an MST over them. Of course, It takes the leaves in the resulting tree, some of the are no longer leaves. This problem is corrected by a transformation that ?pushes? a leaf down the tree, duplicating its model; this transformation preserves the weight (score) of the tree. By using only , the algorithm simply ?throws away? the entire structure of the previous tree. However, we can also construct new MSTs built from all nodes of the previous tree. For all nodes for 98 :8 which end up as internal nodes, we perform the same transformation described above. In both cases, this transformation is not unique, as it depends on the order in which the steps are executed; see Fig. 1(b). The algorithm therefore generates an entire pool of  #  #  #   #                       #  #        #    #  ,9#  # '% &1 ( +-, 21                    candidate trees (from both and ), generated using different random resolutions of ambiguities in the weight-preserving transformation. For each such tree, the CPM learning algorithm is used to find an optimal setting of the parameters. The trees are evaluated relative to our score ( ), and the highest scoring tree is kept. The tree just constructed has a new set of CPMs, so we can repeat this process. To detect termination, the algorithm also keeps the tree from the previous iteration, and terminates when the score of all trees in the newly constructed pool is lower than the score of the best tree from previous iteration. Finally, we address the fact that the data we have is incomplete, in that the assignments of data instances to classes is not determined. We address the problem of incomplete data using the standard Expectation Maximization (EM) algorithm [2] and the structural EM algorithm [4] which extends EM to the problem of model selection. Starting from an initial model, the algorithm iterates the following two steps: The E-step computes the distribution over the unobserved variables given the observed data and the current model. In our case, the distribution over the unobserved variables is computed by evalu8 ating for all 8 . The M-step learns new models that increase the expected log likelihood of the data, relative to the distribution computed in the E-step. In our case, the M-step is precisely the algorithm for complete data described above, but using a soft assignment of data instances to nodes in the tree. The full algorithm is shown in Fig. 2. A simple analysis along the lines of [4] can be used to show that the log-probability increases at every M-step. We therefore obtain the following theorem: %'&( +-,. / " 1 & ." 0 +-,.& . " 0  2/ ,/. " 0 2 1 %'&( " 3/ ")/ +-,.5/<" 1 Theorem 3.1 The algorithm in Fig. 2 converges to a local maximum of 1. Initialize J D = @ 3 5 5 5 3 2. Repeat until convergence: %*&+( +-, /<" 1 . @  B and the models at the leaves. Randomly initialize  . 3    M =@ N @ -B -step: 3 i. Choose an MST over some subset of 3 5 5 5 3 , using as edge weights. ii. Transform the MST so that 3 5 5 5 3 become leaves. 3 5 5 5 3 (b) -step: For , compute the posterior probabilities for the indicator variable   . For : (a) D    = !D B 9 DF; = D > @ -step: Update the CPMs and . Let =  B  N 9D  "  = # 9 D $ % & $ ' IE(   (c) 3   3     -     JLBF; = >-@ NBN &%'  45 D  !" . Then: D B ; = =  B # >) B %'  3   3  3 -  ,5 Figure 2: Abstraction Hierarchy Learning Algorithm 4 Experimental Results We focus our experimental results on genomic expression data, although we also provide some results on a text dataset. In gene expression data, the level of mRNA transcript of every gene in the cell is measured simultaneously, using DNA microarray technology. This genomic expression data provides researchers with much insight towards understanding the overall cellular behavior. The most commonly used method for analyzing this data is clustering, a process which identifies clusters of genes that share similar expression patterns (e.g., [3]), and which are therefore also often involved in similar cellular processes.  , in which case We apply PAH to this data, using CPMs of the form KL-distance is simply: IDKL , which is simply the sum of ,.#  >#  1  35476   +* ,- ,  / .102 1 #  89 , - 8  - 8 1 0 7,.#  # 1  ,9#   #  1 squared distances between the means of the corresponding Gaussian components, normalized by their variance. We therefore define IDKL . The most popular clustering method for genomic expression data to date is hierarchical agglomerative clustering (HAC) [3], which builds a hierarchy among the genes by iteratively merging the closest genes relative to some distance metric. We use the same distance metric for HAC. (Note that in HAC the metric is used as the distance between data cases whereas in our algorithm it is used as the distance between models.) To perform a direct comparison between PAH and HAC, we often need to obtain a probabilistic model from HAC. To do so, we create CPMs from the genes that HAC assigned to each internal node. In both PAH and HAC, we then assign each gene (in the training set or the test set) to the hierarchy by choosing the best (highest likelihood) CPM among all the nodes in the   tree (including internal nodes) and recording the probability  that this CPM assigns to the gene. Structure Recovery. A good algorithm for learning abstraction hierarchies should recover the true hierarchy as well as possible. To test this, we generated a synthetic data set, and measured the ability of each method to recover the distances between pairs of instances (genes) in the generating model, where distance here is the length of the path between two genes in the hierarchy. We generated the data set by sampling from the leaves of a PAH; to make the data realistic, we sampled from a PAH that we learned from a real gene expression data set. To allow a comparison with HAC, we generated one data instance from each leaf. We generated data for 80 (imaginary) genes and 100 experiments, for a total of 8000 measurements. For robustness, we generated 5 different such data sets and ran PAH and HAC for each data set. We used the correlation and the error between the pairwise distances in the original    for PAH, and the learned tree as measures of similiarity. The correlation was      for HAC. The average  compared to a much worse error was     for HAC. These results show that PAH recovers an abstraction for PAH and hierarchy much better than HAC. Generalization. We next tested the ability of the different methods to generalize to unobserved (test) data, measuring the extent to which each method captures the underlying structure in the data. We ran these tests on the yeast data set of [6]. We selected 953 genes with significant changes in expression, using their full set of 93 experiments. Again, we ran PAH and HAC and evaluated performance using 5 fold cross validation. For PAH we also used different settings for (the coefficient of the penalty term in ), ) and greatly which explores the performance in the range of only fitting the data ( favoring hierarchies in which nearby models are similar (large ). In both cases, we learned a model using training data, and evaluated the log-likelihood of test instances as described above. The results, summarized in Fig. 3(a), clearly show that PAH generalizes much better to previously unobserved data than HAC and that PAH works best at some tradeoff between fitting the data and generating a hierarchy in which nearby models are similar. Robustness. Our goal in constructing a hierarchy is to extract meaningful biological conclusions from the data. However, data is invariably partial and noisy. If our analysis produces very different results for slightly different training data, the biological conclusions are unlikely to be meaningful. Thus, we want genes that are assigned to nearby nodes in the tree, to be close together also in hierarchies learned from perturbed data sets. We tested robustness to noise by learning a model from the original data set and from perturbed data sets in which we permuted a varying percentage of the expression measuments. We then compared the distances (the path length in the tree) between the nodes assigned to every pair of genes in trees learned from the original data and trees learned from perturbed data sets. The results are shown in Fig. 3(b), demonstrating that PAH pre serves the pairwise distances extremely well even when  of the data is perturbed (and performs reasonably well for    permutation), while HAC completely deteriorates when  of the data is permuted. +-, /#          +   1     :    +-, 21 -90 1 PAH HAC 0.8 Correlationcoefficient Average log probability -92 0.9 -94 -96 -98 -100 0.7 PAH HAC 0.6 0.5 0.4 0.3 0.2 -102 0.1 -104 0 0 2 4 6 8 10 12 14 16 18 20 Lambda 0 20 40 60 80 100 Datapermuted(% ) (a) (b) Figure 3: (a) Generalization to test data (b) Robustness to noise methodolog sigmoid feedforward spline mont nonLinear carlo mcmc causal causal sample vector gradient Model PAH HAC PAH HAC PAH HAC p 90% 80% 70% Training set avg. L1 difference 5  5   5  5  5   5   5   5  5  5   5  5       Test set avg. L1 difference 5   5  5   5 5   5   5    5   5  5   5   5            boltzmann machin causat causat influenc influenc featur Em maxim algorithm counterfactu pearl train maximum likelihood bayesian network method markov forward global variant hidden learn artificial intellig hmm graphic model statist graph parameter acyclic (a) (b) Figure 4: (a) Robustness of PAH and HAC to different subsets of training instances. (b) Word hierarchy learned on Cora data. A second important test is robustness to our particular choice of training data: a particular training set reflects only a subset of the experiments that we could have performed. In this experiment, we used the Yeast Compendium data of [9], which measures the expression profiles triggered by specific gene mutations. We selected 450 genes and all 298 arrays, focusing on genes that changed significantly. For each of three values of  ranging from  to   , we generated ten different training sets by sampling (without replacement)  percent of the 450 genes, the rest of which form a test set. We then placed both training and test genes within the hierarchy. For each data set, every pair of genes either appear together in the training set, the test set, or do not appear together (i.e., one appears in the training set and the other in the test set). We compared, for each pair of genes, their distances in training sets in which they appear together and their distances in test sets in which they appear together. The results are summarized in Fig. 4(a). Our results on the training data show that PAH consistently constructs very similar hierarchies, even from very different subsets of the data. By contrast, the hierarchies constructed by HAC are much less consistent. The results on the test data are even more striking. PAH is very consistent about its classification into the hierachy even of test instances ? ones not used to construct the hierarchy. In fact, there is no significant difference between its performance on the training data and the test data. By contrast, HAC places test instances in very different configurations in different trees, reducing our confidence in the biological validity of the learned structure. Intuitiveness. To get qualitative insight into the hierarchies produced, we ran PAH on 350 documents from the Probabilistic Methods category in the Cora dataset (cora.whizbang.com) and learned hierarchies among the (stemmed) words. We constructed a vector for each word with an entry for each document whose value is the TFIDF-  + weighted frequency of the word within the document. Fig. 4(b) shows parts of the learned hierarchy, consisting of 441 nodes, where we list high confidence words for each node. PAH organized related words into the same region of the tree. Within each region, many words were arranged in a way which is consistent with our intuitive notion of abstraction. 5 Discussion We presented probabilistic abstraction hierarchies, a general framework for learning abstraction hierarchies from data, which relates different classes in the hierarchy by a tree whose nodes correspond to class-specific probability models (CPMs). We utilize a Bayesian approach, where the prior favors hierarchies in which nearby classes have similar data distributions, by penalizing the distance between neighboring CPMs. A unique feature of PAH is the use of global optimization steps for constructing the hierarchy and for finding the optimal setting of the entire set of parameters. This feature differentiates us from many other approaches that build hierarchies by local improvements of the objective function or approaches that optimize a fixed hierarchy [7]. The global optimization steps help in avoiding local maxima and in reducing sensitivity to noise. Our approach leads naturally to a form of parameter smoothing, and provides much better generalization for test data and robustness to noise than other clustering approaches. In principle, we can use any probabilistic model for the CPM as long as it defines a probability distribution over the state space. We have recently [14] applied this approach to the substantially more complex problem of clustering proteins based on their amino acid sequence using profile HMMs [11]. Acknowledgements. We thank Nir Friedman for useful comments. This work was supported by NSF Grant ACI-0082554 under the NSF ITR program, and by the Sloan Foundation. Eran Segal was also supported by a Stanford Graduate Fellowship (SGF). References [1] P. Cheeseman and J. Stutz. Bayesian Classification (AutoClass): Theory and Results. AAAI Press, 1995. [2] A. P. Dempster, N. M. Laird, and D. B. Rubin. Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society, B 39:1?39, 1977. [3] M. Eisen, P. Spellman, P. Brown, and D. Botstein. Cluster analysis and display of genome-wide expression patterns. PNAS, 95:14863?68, 1998. [4] N. Friedman. The Bayesian structural EM algorithm. In Proc. UAI, 1998. [5] N. Friedman, M. Ninio, I. Pe?er, and T. Pupko. A structural EM algorithm for phylogentic inference. In Proc. RECOMB, 2001. [6] A.P. Gasch et al. Genomic expression program in the response of yeast cells to environmental changes. Mol. Bio. Cell, 11:4241?4257, 2000. [7] T. Hofmann. The cluster-abstraction model: Unsupervised learning of topic hierarchies from text data. In Proc. IJCAI, 1999. [8] T. Hofmann. The cluster-abstraction model: Unsupervised learning of topic hierarchies from text data. In Proc. International Joint Conference on Artificial Intelligence, 1999. [9] T. R. Hughes et al. Functional discovery via a compendium of expression profiles. Cell, 102(1):109?26, 2000. [10] F.K. Hwang, D.S.Richards, and P. Winter. The Steiner Tree Problem. Annals of Discrete Mathematics, Vol. 53, North-Holland, 1992. [11] A. Krogh, M. Brown, S. Mian, K. Sjolander, and D. Haussler. Hidden markov models in computational biology: Applications to protein modeling. Mol. Biology, 235:1501?1531, 1994. [12] A. McCallum, R. Rosenfeld, T. Mitchell, and A. Ng. Improving text classification by shrinkage in a hierarchy of classes. In Proc. ICML, 1998. [13] M. Meila and M.I. Jordan. Learning with mixtures of trees. Machine Learning, 1:1?48, 2000. [14] E. Segal and D. Koller. Probabilistic hierarchical clustering for biological data. In RECOMB, 2002.
2065 |@word hierachy:1 sgf:1 termination:1 simplifying:1 thereby:1 initial:2 configuration:1 series:1 score:4 selecting:2 denoting:1 document:4 imaginary:1 steiner:3 current:1 com:1 stemmed:1 must:2 mst:5 additive:1 numerical:1 realistic:1 hofmann:2 designed:1 update:1 generative:1 discovering:1 leaf:28 item:3 parameterization:1 selected:2 intelligence:1 mccallum:1 provides:3 iterates:1 node:22 daphne:1 mathematical:1 along:2 constructed:5 direct:1 become:1 qualitative:1 fitting:2 g4:5 pairwise:2 expected:2 indeed:1 behavior:1 begin:1 unrelated:1 underlying:1 maximizes:1 lowest:1 substantially:2 finding:3 transformation:6 unobserved:4 duplicating:1 every:4 exactly:1 decouples:1 bio:1 grant:1 appear:4 local:11 modify:1 tends:1 analyzing:2 path:2 might:1 chose:1 conversely:1 hmms:1 range:2 graduate:1 unique:3 practice:2 hughes:1 procedure:5 significantly:1 whizbang:1 pre:1 word:7 confidence:2 protein:2 get:2 close:4 selection:2 restriction:2 optimize:3 maximizing:1 straightforward:1 attention:1 starting:1 independently:2 convex:1 focused:1 resolution:1 mrna:1 recovery:1 assigns:1 pure:1 m2:4 insight:2 haussler:1 array:1 pull:1 notion:2 pupko:1 annals:1 hierarchy:51 machin:1 element:1 intuitiveness:1 particularly:1 richards:1 featur:1 bottom:2 role:1 subproblem:1 observed:1 capture:1 region:2 highest:2 ran:4 principled:2 intuition:1 mentioned:1 convexity:3 dempster:1 idkl:5 completely:1 easily:1 joint:5 train:1 artificial:2 formation:1 choosing:1 whose:4 heuristic:2 stanford:7 larger:2 solve:1 ability:2 favor:1 g1:5 rosenfeld:1 transform:1 itself:1 noisy:1 laird:1 advantage:2 triggered:1 sequence:1 propose:2 neighboring:2 ramification:1 date:1 iff:1 achieve:1 intuitive:2 parent:1 cluster:7 optimum:1 convergence:2 empty:1 produce:2 generating:2 ijcai:1 converges:2 help:2 fixing:1 measured:3 transcript:1 krogh:1 throw:1 c:3 trading:1 closely:1 require:3 assign:1 generalization:3 tfidf:1 biological:5 hold:2 around:1 considered:1 substituting:1 estimation:1 proc:5 applicable:1 cpm:19 propensity:1 sensitive:1 bridge:1 create:1 reflects:1 weighted:1 cora:3 clearly:1 genomic:4 gaussian:4 always:1 avoid:1 shrinkage:2 varying:1 probabilistically:1 focus:1 improvement:2 consistently:1 likelihood:8 greatly:1 contrast:2 sense:1 detect:1 inference:1 abstraction:19 typically:2 entire:4 unlikely:1 hidden:4 koller:3 relation:1 favoring:1 semantics:1 overall:2 issue:1 among:3 classification:3 augment:1 denoted:1 sjolander:1 smoothing:2 special:1 summed:1 fairly:2 initialize:2 construct:4 ng:1 sampling:2 identical:1 represents:2 biology:2 unsupervised:2 icml:1 peaked:2 np:1 spline:1 simplify:1 randomly:2 winter:1 simultaneously:2 preserve:1 individual:1 m4:7 consisting:1 replacement:1 friedman:3 invariably:1 mixture:4 edge:3 closer:2 partial:1 stutz:1 tree:58 incomplete:3 divide:1 euclidean:1 penalizes:1 desired:1 causal:2 theoretical:1 instance:22 soft:1 modeling:1 measuring:1 assignment:6 maximization:1 ating:1 decomposability:1 subset:4 entry:1 graphic:1 perturbed:4 synthetic:3 density:1 peak:1 sensitivity:2 explores:1 ie:1 international:1 probabilistic:16 off:1 pool:2 together:5 autoclass:2 squared:2 ambiguity:1 again:1 aaai:1 containing:1 choose:1 worse:1 lambda:1 creating:3 segal:3 summarized:2 north:1 coefficient:1 sloan:1 depends:1 performed:1 root:1 try:1 start:1 recover:2 mutation:1 minimize:1 variance:1 characteristic:1 acid:1 correspond:2 generalize:2 raw:1 bayesian:6 produced:2 basically:1 cpms:30 carlo:1 researcher:1 definition:3 frequency:1 involved:1 obvious:1 naturally:4 associated:6 recovers:1 sampled:3 newly:1 pah:35 dataset:2 popular:1 mitchell:1 knowledge:2 organized:3 focusing:1 appears:1 follow:1 botstein:1 response:1 arranged:1 evaluated:3 just:1 mian:1 until:2 correlation:2 nonlinear:1 defines:4 perhaps:1 yeast:3 hwang:1 validity:1 normalized:1 true:1 brown:2 hence:2 assigned:4 symmetric:1 iteratively:1 round:1 m5:3 complete:2 performs:1 l1:2 percent:1 ranging:1 novel:1 recently:1 common:2 sigmoid:1 specialized:1 permuted:2 multinomial:4 functional:1 discussed:1 interpretation:1 m1:4 measurement:1 significant:2 rd:1 meila:1 mathematics:1 similarly:1 specification:1 similarity:1 longer:1 posterior:1 closest:1 optimizing:1 optimizes:1 belongs:1 arbitrarily:1 scoring:1 preserving:2 minimum:2 additional:1 care:1 somewhat:2 converge:2 ii:1 relates:1 multiple:1 full:2 pnas:1 reduces:2 cross:1 long:1 post:1 variant:3 involving:1 essentially:1 expectation:1 df:1 metric:3 iteration:4 cell:4 whereas:1 want:1 fellowship:1 microarray:1 rest:2 hac:23 comment:1 induced:1 tend:1 virtually:1 undirected:2 thing:1 subject:1 recording:1 jordan:1 structural:5 feedforward:1 m6:3 variety:1 fit:1 itr:1 tradeoff:1 expression:16 penalty:2 suffer:1 proceed:1 useful:2 locally:1 statist:1 ten:1 category:1 dna:1 generate:2 percentage:1 nsf:2 deteriorates:1 discrete:1 vol:1 key:2 demonstrating:1 penalizing:1 kept:1 utilize:1 graph:1 sum:2 striking:1 place:2 extends:2 utilizes:3 prefer:1 guaranteed:3 tackled:1 display:1 fold:1 precisely:2 flat:1 nearby:10 generates:1 aspect:1 compendium:2 extremely:1 according:1 terminates:1 slightly:1 em:8 g3:5 making:1 lem:1 intellig:1 taken:1 visualization:1 remains:1 previously:1 discus:1 turn:2 differentiates:1 end:1 serf:1 generalizes:1 apply:1 hierarchical:7 away:2 enforce:2 robustness:7 substitute:1 original:3 top:1 clustering:12 include:1 ensure:1 remaining:1 exploit:1 build:2 conquer:1 society:1 objective:1 eran:3 gradient:1 distance:23 separate:1 thank:1 hmm:1 topic:2 agglomerative:4 extent:2 cellular:2 gasch:1 spanning:1 enforcing:1 length:2 equivalently:1 executed:1 taxonomy:6 subproblems:2 negative:1 proper:1 boltzmann:1 unknown:2 perform:2 observation:1 markov:3 defining:1 unconnected:1 dirk:1 pair:5 kl:1 learned:12 pearl:1 address:2 pattern:2 program:2 built:1 including:1 royal:1 difficulty:1 examination:1 ormoneit:2 indicator:1 cheeseman:1 spellman:1 technology:1 identifies:1 extract:1 nir:1 text:7 prior:8 understanding:2 acknowledgement:1 discovery:1 relative:6 permutation:1 limitation:1 recomb:2 acyclic:1 validation:1 foundation:1 consistent:3 rubin:1 principle:2 share:1 prone:1 course:1 penalized:1 changed:1 surprisingly:1 last:1 repeat:2 mont:1 placed:1 supported:2 allow:1 wide:3 neighbor:2 volves:1 dimension:1 genome:1 computes:1 eisen:1 forward:1 commonly:1 avg:2 gene:26 overcomes:1 keep:2 global:11 uai:1 assumed:1 aci:1 continuous:1 iterative:2 robin:1 learn:4 reasonably:1 robust:1 sem:3 mol:2 improving:1 complex:2 constructing:4 domain:8 noise:7 profile:3 child:1 repeated:1 amino:1 fig:8 fashion:1 similiarity:1 candidate:1 pe:1 learns:1 down:2 theorem:2 specific:4 showing:1 er:1 list:1 exists:1 merging:1 effectively:1 maxim:1 subtree:1 illustrates:1 push:1 simply:4 desire:1 g2:10 holland:1 environmental:1 conditional:1 goal:6 careful:1 towards:1 hard:1 change:2 specifically:1 typical:1 reducing:3 corrected:1 acting:1 determined:1 total:1 experimental:3 m3:6 meaningful:2 rarely:1 internal:5 incorporate:1 dept:3 mcmc:1 tested:2 avoiding:1
1,169
2,066
Rao-Blackwellised Particle Filtering Data Augmentation . VIa Christophe Andrieu N ando de Freitas Arnaud Doucet Statistics Group University of Bristol University Walk Bristol BS8 1TW, UK Computer Science UC Berkeley 387 Soda Hall, Berkeley CA 94720-1776, USA EE Engineering University of Melbourne Parkville, Victoria 3052 Australia C.Andrieu @bristol.ac.uk jfgf@cs.berkeley.edu doucet@ee .mu.oz.au Abstract In this paper, we extend the Rao-Blackwellised particle filtering method to more complex hybrid models consisting of Gaussian latent variables and discrete observations. This is accomplished by augmenting the models with artificial variables that enable us to apply Rao-Blackwellisation. Other improvements include the design of an optimal importance proposal distribution and being able to swap the sampling an selection steps to handle outliers. We focus on sequential binary classifiers that consist of linear combinations of basis functions , whose coefficients evolve according to a Gaussian smoothness prior. Our results show significant improvements. 1 Introduction Sequential Monte Carlo (SMC) particle methods go back to the first publically available paper in the modern field of Monte Carlo simulation (Metropolis and Ulam 1949) ; see (Doucet, de Freitas and Gordon 2001) for a comprehensive review. SMC is often referred to as particle filtering (PF) in the context of computing filtering distributions for statistical inference and learning. It is known that the performance of PF often deteriorates in high-dimensional state spaces. In the past, we have shown that if a model admits partial analytical tractability, it is possible to combine PF with exact algorithms (Kalman filters, HMM filters , junction tree algorithm) to obtain efficient high dimensional filters (Doucet, de Freitas, Murphy and Russell 2000, Doucet, Godsill and Andrieu 2000). In particular, we exploited a marginalisation technique known as Rao-Blackwellisation (RB). Here, we attack a more complex model that does not admit immediate analytical tractability. This probabilistic model consists of Gaussian latent variables and binary observations. We show that by augmenting the model with artificial variables, it becomes possible to apply Rao-Blackwellisation and optimal sampling strategies. We focus on the problem of sequential binary classification (that is, when the data arrives one-at-a-time) using generic classifiers that consist of linear combinations of basis functions, whose coefficients evolve according to a Gaussian smoothness prior (Kitagawa and Gersch 1996). We have previously addressed this problem in the context of sequential fault detection in marine diesel engines (H0jen-S0rensen, de Freitas and Fog 2000). This application is of great importance as early detection of incipient faults can improve safety and efficiency, as well as, help to reduce downtime and plant maintenance in many industrial and transportation environments. 2 Model Specification and Estimation Objectives Let us consider the following binary classification model. Given at time t = 1,2, .. . an input Xt we observe Zt E {O, I} such that Pr( Zt = llxt ,.8t ) = CP(f(xl, .8t}), (1) vk J::oo where CP (u) = exp (_a 2 /2) da is the cumulative function of the standard normal distribution. This is the so-called pro bit link. By convention, researchers tend to adopt a logistic (sigmoidal) link function 'P (u) = (1 + exp (_U)) -1 . However, from a Bayesian computational point of view, the probit link has many advantages and is equally valid. The unknown function is modeled as K !(Xt, .8t) = L .8t,k\[ldxt) = \[IT (Xt).8t , k=1 where we have assumed that the basis functions \[I (Xt) ? (\[11 (Xt) , ... , \[I K (Xt)/ do not depend on unknown parameters; see (Andrieu, de Freitas and Doucet 1999) for the more general case . .8t ? (.8t,1,' .. ,.8t,K )T E ~K is a set of unknown time-varying regression coefficients. To complete the model , we assume that they satisfy .8t = At.8t-1 + BtVt, .80'" N (rna, Po) (2) where Vt i?~.:f N (0 , In.) and A and B control model correlations and smoothing (regularisation). Typically K is rather large, say 10 or 100, and the bases \[Ik (.) are multivariate splines, wavelets or radial basis functions (Holmes and Mallick 1998). 2.1 Augmented Statistical Model We augment the probabilistic model artificially to obtain more efficient sampling algorithms, as will be detailed in the next section. In particular, we introduce the set of independent variables Yt , such that Yt =! (Xt,.8t) + nt, (3) h i.i.d. N (0 1) d d fi were nt '" "an e ne Zt that one has Pr ( Zt = 11 Xt, .8t ) = = {I0 if Yt > 0, otherwise. It is then easy to check CP (f (Xl, .8t)) . This data augmentation strategy was first introduced in econometrics by economics Nobel laureate Daniel McFadden (McFadden 1989). In the MCMC context, it has been used to design efficient samplers (Albert and Chib 1993). Here, we will show how to take advantage of it in an SMC setting. 2.2 Estimation objectives Given, at time t , the observations Ol:t ? (Xl:t, Zl:t), any Bayesian inference is based on the posterior distribution 1 P (d.8o:tl Ol:t)' We are, therefore, interested in estimating sequentially in time this distribution and some of its features , such as IFor any B, we use P (dBo,tl au) to denote the distribution and p (Bo,tl au) to denote the density, where P (dBo,tl au) = p (Bo,tl au) dBo,t. Also, Bo,t ~ {Bo, BI , ... , Bd . lE ( f (xt, ,Bt) I Ol:t) or the marginal predictive distribution at time t for new input data Xt+1, that is Pr (Zt+1 = 11 01:t, xHd. The posterior density satisfies a time recursion according to Bayes rule, but it does not admit an analytical expression and, consequently, we need to resort to numerical methods to approximate it. 3 Sequential Bayesian Estimation via Particle Filtering A straightforward application of SMC methods to the model (1)-(2) would focus on sampling from the high-dimensional distribution P (d,Bo:t I 01:t) (H0jen-S0rensen et al. 2000). A substantially more efficient strategy is to exploit the augmentation of the model to sample only from the low-dimensional distribution P ( dY1:t I01:t). The low-dimensional samples allow us then to compute the remaining estimates analytically, as shown in the following subsection. 3.1 Augmentation and Rao-Blackwellisation Consider the extended model defined by equations (1)-(2)-(3). One has p(,Bo:tlo1:t) = J p( ,Bo:tl x 1:t,Y1:t)p(Y1:tl o1:t)dY1:t? Thus if we have a Monte Carlo approximation of P (dY1:t I ol:d of the form then P (,Bo:tl 01:t) can be approximated via N PN (,Bo:tl 01:t) = L w~i)p ( ,Bo:tl x1:t,yi:O ' i=l that is a mixture of Gaussians. From this approximation, one can estimate lE(,Btlxl:t,Yl:t) and lE(,Bt-Llxl:t,Yl:t). For example, an estimate of the predictive distribution is given by PrN(Zt+1 = Il ol:t,XH1) = J Pr( Zt+1 = lIYH1)PN(dYl:t+1 lol:t, xt+1) (4) N ,, (i) ) = ~ W t(i) ][(0,+00) ( YHl , i=l where Y~21 ~ P ( dYHll Xl:t+1, Yi~~). This shows that we can restrict ourselves to the estimation of P (Y1:t1 Ol:t) for inference purposes. In the SMC framework, we must estimate the "target" density P (Y1:t1 Ol:t) pointwise up to a normalizing constant. By standard factorisation, one has t p(Yl:tlol:t) IT Pr( zk IYk)p(Ykl xl:k,Yl:k-l), wherep(YIIY1:0,Xl:0) ,@,p(Yll xd? k=l Since Pr (Zk I Yk) is known, we only need to estimate P (Yk I Xl:k, Yl:k-d up to a norIX malizing constant. This predictive density can be computed using the Kalman filter. Given (Xl:k' Yl:k-l), the Kalman filter equations are the following. Set ,Bo lo = mo and ~o l o = ~o, then for t = 1, ... , k - 1 compute ,Bt lt-1 = At,Bt- 1It- 1 + BtBI (xt} ~tlt - 1 \[I (Xt) + 1 Yt lt-1 = \[IT (Xt) ,Bt lt-1 ,Btlt = ,Bt lt -1 + ~tlt- 1 \[I (Xt) St- 1 (Yt - Yt lt - t) ~tit = ~t l t-1 - ~t l t-1 \[I (xt} St- 1\[lT (Xt) ~t l t-1' ~t l t-1 = At~t-1 I t-1AI St = \[IT (5) where ,Bt lt - 1 ~ 1E(,BtIXl:t-1,Yl:t-d, ,Btlt ~ 1E(,Btlxl:t,Yl:t), Ytlt - 1 IE(Ytlxl:t,Yl:t - d, ~t l t-1 ~ cov(,BtIXl:t- 1,Y1:t- 1), ~t lt ~ cov(,Btlxl:t,Yl:t) and St ~ cov (Ytl Xl:t,Y1:t-1). One obtains P (Yk I X1:k, Y1:k-d = 3.2 (6) N (Yk;Y klk- 1' Sk) . Sampling Algorithm In this section, we briefly outline the PF algorithm for generating samples from p(dYl:tlol:t). (For details, please refer to our extended technical report at http://www . cs. berkeley. edu/ '" jfgf /publications . html.) Assume that at time t - 1 we have N particles {Yi~Ld~l distributed according to P (dYl:t - 11 ol:t- d from which one can get the following empirical distribution approximation 1 PN (dYl:t-11 ol:t-d =N N L JYi~;_l (dYl:t-d . i= l Various SMC methods can be used to obtain N new paths {Yi~~}~l distributed approximately according to P (dYl:t1 Ol:t)' The most successful of these methods typically combine importance sampling and a selection scheme. Their asymptotic convergence (N --t 00) is satisfied under mild conditions (Crisan and Doucet 2000). Since the selection step is standard (Doucet et al. 2001), we shall concentrate on describing the importance sampling step. To obtain samples from P( dYl:t IOl:t), we can sample from a proposal distribution Q(dYl:t) and weight the samples appropriately. Typically, researchers use the transition prior as proposal distribution (Isard and Blake 1996). Here, we implement an optimal proposal distribution, that is one that minimizes the variance of the importance weights W (Yl:t) conditional upon not modifying the path Y1:t-1' In our case, we have ( I ) P Yt X1:t,Yl:t-1,Zt ex: {p(YtIXl:t ,Y1:t-dlI[o,+ oo) (Yt) p(Ytlxl:t ,Yl:t-dlI(- oo,o) (Yt) if Zt = 1 if Zt = 0 ' which is a truncated Gaussian version of (6) of and consequently W (Yl:t) ex: Pr (Zt I Xl:t, Y1:t - d = (1 _ <I> ( _ Y$,l ) ) z, <I> ( _ Y$,l ) 1-z, (7) The algorithm is shown in Figure 1. (Please refer to our technical report for convergence details.) Remark 1 When we adopt the optimal proposal distribution, the importance weight Wt ex: Pr (Zt I X1:t, Y1:t - d does not depend on Yt. It is thus possible to carry out the selection step before the sampling step. The algorithm is then similar to the auxiliary variable particle filter of (Pitt and Shephard 1999). This modification to the original algorithm has important implications. It enables us to search for more Sequential importance sampling step ? For . t ? For i -(i) h. (i) :::{i) = 1, ... , N, (3t lt-1 = (3t lt-1 and sample Yt = 1, ... , N, ~ P ( (i) ) dYtl Xl:t, Yl:t-1 ' Zt . evaluate the importance weights using (7). Selection step ? Multiply/Discard particles {~i ),,B~i~ _l}~l with respect to high/low impor. hts W (i) to 0 b ' N partlc . Ies { Yt(i) , (3(i) }N ? tance welg tam t lt- 1 i=l ' t Updatmg step ? Compute ~t+1 I t given ~t l t - 1' ? For i = 1, ... , N, C) use one step of the Kalman recursion (5) to compute {,B~i~ l l t } -C) given {y/ ,(3 ti t-1 } and ~t l t-1' Figure 1: RBPF for semiparametric binary classification. likely regions of the posterior at time t-1 using the information at time t to generate better samples at time t. In practice, this increases the robustness of the algorithm to outliers and allows us to apply it in situations where the distributions are very peaked (e.g., econometrics and almost deterministic sensors and actuators). Remark 2 Th e covariance updates of the Kalman jilter are outside the loop over particles. This results in substantial computational savings. 4 Simulations To compare our model , using the RBPF algorithm, to standard logistic and probit classification with PF, we generated data from clusters that change with time as shown in Figure 2. This data set captures the characteristics of a fault detection problem that we are currently studying. (For some results of applying PF to fault detection in marine diesel engines, please refer to (H0jen-S0rensen et al. 2000). More results will become available once permission is granted.) This data cannot be easily separated with an algorithm based on a time-invariant model. For the results presented here, we set the initial distributions to: (30 '" N(O , 51) and Yo '" N(O, 51). The process matrices were set to A = I and B = JI, where 82 = 0.1 is a smoothing parameter. The number of bases (cubic splines with random locations) was set to 10. (It is of course possible, when we have some data already, to initialise the bases locations so that they correspond to the input data. This trick for efficient classification in high dimensional input spaces is used in the support vector machines setting (Vapnik 1995).) The experiment was repeated with the number of particles varying between 10 and 400. Figure 3 shows the "value for money" summary plot. The new algorithm has a lower computational cost and shows a significant reduction in estimation variance. Note that the computation of the RBPF stays consistently low even for small numbers of particles. This has enabled us to apply the technique to large models consisting of hundreds of Bases using a suitable regulariser. Another advantage of PF algorithms for classification is that they yield entire probability estimates of class membership as shown in Figure 4. 0'<>0 -:5' - - - -o:--- ---::5 -:5: ---'-'-- -0 : ------::5 Data from t=1 to t=100 Data from t=1 00 to t=200 -5'---------5 0 - 5 '-------"----- 5 -5 Data from t=200 to t=300 0 5 Data from t=1 to t=300 Figure 2: Time-varying data. 35 !... . 'j r,. I I ' . ~ , ...... ... . I i T L - ... - I -n. ... "',i ... T L~ . ? ,- 1''''' ~ ... ~.......... i -,... . 50'----~--L-~--~8-~,0~-~,2--,~ 4-~,6-~,8 Computation (flops) ,10' Figure 3: Number of classification errors as the number of particles varies between 10 and 400 (different computational costs). The algorithm with the augmentation trick (RBPF) is more efficient than standard PF algorithms. 5 Conclusions In this paper, we proposed a dynamic Bayesian model for time-varying binary classification and an efficient particle filtering algorithm to perform the required computations. The efficiency of our algorithm is a result of data augmentation, RaoBlackwellisation, adopting the optimal importance distribution, being able to swap the sampling and selection steps and only needing to update the Kalman filter means in the particles loop. This extends the realm of efficient particle filtering to the ubiquitous setting of Gaussian latent variables and binary observations. Extensions to n-ary observations, different link functions and estimation of the hyper-parameters can be carried out in the same framework. 50 -1 Figure 4: Predictive density. References Albert , J. and Chib , S. (1993) . Bayesian analysis of binary and polychotomous response data, Journal of the American Statistical Association 88(422): 669- 679. Andrieu, C., de Freitas, N. and Doucet , A. (1999) . Sequential Bayesian estimation and model selection applied to neural networks, Technical Report CUED/F-INFENG/TR 341, Cambridge University Engineering Department. Crisan, D. and Doucet, A. (2000) . Convergence of sequential Monte Carlo methods, Technical Report CUED/F-INFENG/TR 381, Cambridge University Engineering Department. Doucet , A. , de Freitas, N. and Gordon, N. J. (eds) (2001) . Sequential Monte Carlo Methods in Practice, Springer-Verlag. Doucet , A. , de Freitas, N. , Murphy, K. and Russell, S. (2000). Rao blackwellised particle filtering for dynamic Bayesian networks, in C. Boutilier and M. Godszmidt (eds), Uncertainty in Artificial Intelligence, Morgan Kaufmann Publishers, pp . 176- 183. Doucet, A., Godsill, S. and Andrieu, C. (2000). On sequential Monte Carlo sampling methods for Bayesian filtering , Statistics and Computing 10(3): 197- 208. H0jen-S0rensen, P. , de Freitas, N. and Fog, T. (2000). On-line probabilistic classification with particle filters, IEEE N eural Networks for Signal Processing, Sydney, Australia. Holmes, C. C. and Mallick, B. K. (1998). Bayesian radial basis functions of variable dimension, Neural Computation 10(5): 1217- 1233. Isard, M. and Blake, A. (1996) . Contour tracking by stochastic propagation of conditional density, European Conference on Computer Vision, Cambridge, UK, pp. 343- 356. Kitagawa, G. and Gersch, W. (1996). Smoothn ess Priors Analysis of Tim e Series, Vol. 116 of Lecture Notes In Statistics, Springer-Verlag. McFadden, D. (1989). A method of simulated momemts for estimation of discrete response models without numerical integration, Econometrica 57: 995- 1026. Metropolis, N. and Uiam, S. (1949). The Monte Carlo method, Journal of th e American Statistical Association 44(247): 335- 341. Pitt , M. K. and Shephard, N. (1999). Filtering via simulation: Auxiliary particle filters , Journal of the American Statistical Association 94(446): 590- 599. Vapnik , V. (1995). Th e Nature of Statistical Learning Th eory, Springer-Verlag, New York.
2066 |@word mild:1 version:1 briefly:1 simulation:3 covariance:1 tr:2 klk:1 carry:1 ld:1 reduction:1 initial:1 series:1 daniel:1 past:1 freitas:9 nt:2 bd:1 must:1 numerical:2 enables:1 plot:1 hts:1 update:2 isard:2 intelligence:1 es:1 marine:2 location:2 attack:1 sigmoidal:1 welg:1 become:1 ik:1 consists:1 combine:2 introduce:1 ol:10 pf:8 becomes:1 estimating:1 substantially:1 minimizes:1 blackwellised:3 berkeley:4 ti:1 xd:1 classifier:2 uk:3 control:1 zl:1 safety:1 t1:3 engineering:3 before:1 path:2 approximately:1 au:5 smc:6 bi:1 practice:2 implement:1 empirical:1 downtime:1 radial:2 get:1 cannot:1 selection:7 context:3 applying:1 www:1 deterministic:1 transportation:1 yt:12 go:1 economics:1 straightforward:1 factorisation:1 rule:1 holmes:2 initialise:1 enabled:1 handle:1 target:1 exact:1 trick:2 approximated:1 econometrics:2 capture:1 region:1 russell:2 prn:1 yk:4 substantial:1 environment:1 mu:1 econometrica:1 dynamic:2 depend:2 tit:1 predictive:4 upon:1 efficiency:2 swap:2 basis:5 po:1 dyl:8 easily:1 various:1 iol:1 separated:1 monte:7 artificial:3 hyper:1 outside:1 whose:2 say:1 otherwise:1 statistic:3 cov:3 advantage:3 analytical:3 loop:2 oz:1 convergence:3 cluster:1 ulam:1 generating:1 help:1 oo:3 cued:2 ac:1 augmenting:2 tim:1 shephard:2 sydney:1 auxiliary:2 c:2 convention:1 concentrate:1 filter:9 modifying:1 stochastic:1 australia:2 enable:1 kitagawa:2 extension:1 hall:1 blake:2 normal:1 exp:2 great:1 mo:1 pitt:2 early:1 adopt:2 purpose:1 estimation:8 currently:1 bs8:1 sensor:1 gaussian:6 rna:1 rather:1 pn:3 crisan:2 varying:4 publication:1 ytl:1 focus:3 yo:1 improvement:2 vk:1 consistently:1 check:1 industrial:1 inference:3 publically:1 i0:1 membership:1 typically:3 bt:7 entire:1 interested:1 classification:9 html:1 augment:1 smoothing:2 integration:1 uc:1 marginal:1 field:1 once:1 saving:1 sampling:11 peaked:1 report:4 spline:2 gordon:2 modern:1 chib:2 comprehensive:1 murphy:2 consisting:2 ourselves:1 ando:1 detection:4 multiply:1 mixture:1 arrives:1 fog:2 implication:1 partial:1 tree:1 impor:1 walk:1 melbourne:1 rao:7 tractability:2 cost:2 hundred:1 successful:1 llxl:1 varies:1 st:4 density:6 ie:1 stay:1 probabilistic:3 yl:15 polychotomous:1 augmentation:6 satisfied:1 ykl:1 admit:2 tam:1 resort:1 american:3 de:9 coefficient:3 satisfy:1 view:1 bayes:1 il:1 variance:2 characteristic:1 kaufmann:1 correspond:1 yield:1 bayesian:9 carlo:7 researcher:2 dbo:3 bristol:3 ary:1 ed:2 i01:1 pp:2 subsection:1 realm:1 ubiquitous:1 back:1 response:2 correlation:1 propagation:1 llxt:1 logistic:2 h0jen:4 usa:1 andrieu:6 analytically:1 arnaud:1 please:3 outline:1 complete:1 cp:3 pro:1 fi:1 dy1:3 ji:1 extend:1 association:3 significant:2 refer:3 cambridge:3 ai:1 smoothness:2 particle:18 lol:1 specification:1 money:1 base:4 multivariate:1 posterior:3 discard:1 verlag:3 binary:8 christophe:1 fault:4 accomplished:1 exploited:1 yi:4 vt:1 morgan:1 signal:1 needing:1 technical:4 jfgf:2 equally:1 y:1 infeng:2 regression:1 maintenance:1 vision:1 albert:2 adopting:1 proposal:5 semiparametric:1 addressed:1 publisher:1 appropriately:1 marginalisation:1 jyi:1 tend:1 ee:2 easy:1 restrict:1 reduce:1 expression:1 granted:1 york:1 remark:2 boutilier:1 detailed:1 ifor:1 eory:1 http:1 generate:1 deteriorates:1 rb:1 discrete:2 shall:1 vol:1 group:1 incipient:1 uncertainty:1 soda:1 extends:1 almost:1 bit:1 department:2 according:5 combination:2 gersch:2 tw:1 metropolis:2 modification:1 outlier:2 invariant:1 pr:8 equation:2 previously:1 describing:1 studying:1 available:2 junction:1 gaussians:1 victoria:1 apply:4 observe:1 actuator:1 generic:1 permission:1 robustness:1 original:1 remaining:1 include:1 exploit:1 objective:2 already:1 strategy:3 link:4 simulated:1 hmm:1 nobel:1 kalman:6 o1:1 modeled:1 pointwise:1 godsill:2 design:2 regulariser:1 zt:13 unknown:3 perform:1 observation:5 truncated:1 immediate:1 situation:1 extended:2 flop:1 y1:11 introduced:1 required:1 dli:2 engine:2 yll:1 able:2 tance:1 mallick:2 suitable:1 parkville:1 hybrid:1 recursion:2 scheme:1 improve:1 iyk:1 ne:1 carried:1 prior:4 review:1 evolve:2 regularisation:1 asymptotic:1 plant:1 probit:2 lecture:1 mcfadden:3 filtering:10 rbpf:4 lo:1 course:1 summary:1 blackwellisation:4 allow:1 distributed:2 dimension:1 valid:1 cumulative:1 transition:1 contour:1 tlt:2 approximate:1 obtains:1 laureate:1 doucet:13 sequentially:1 assumed:1 search:1 latent:3 sk:1 nature:1 zk:2 ca:1 complex:2 artificially:1 european:1 da:1 repeated:1 x1:4 augmented:1 eural:1 referred:1 tl:10 wherep:1 cubic:1 xl:11 wavelet:1 xh1:1 xt:17 admits:1 normalizing:1 consist:2 vapnik:2 sequential:10 importance:9 lt:11 likely:1 tracking:1 bo:11 springer:3 satisfies:1 yhl:1 conditional:2 consequently:2 change:1 sampler:1 wt:1 called:1 s0rensen:4 support:1 evaluate:1 mcmc:1 ex:3
1,170
2,067
Perceptual Metamers in Stereoscopic Vision Benjamin T. Backus* Department of Psychology University of Pennsylvania Philadelphia, PA 19104-6196 backus@psych.upenn.edu Abstract Theories of cue combination suggest the possibility of constructing visual stimuli that evoke different patterns of neural activity in sensory areas of the brain, but that cannot be distinguished by any behavioral measure of perception. Such stimuli, if they exist, would be interesting for two reasons. First, one could know that none of the differences between the stimuli survive past the computations used to build the percepts. Second, it can be difficult to distinguish stimulus-driven components of measured neural activity from top-down components (such as those due to the interestingness of the stimuli). Changing the stimulus without changing the percept could be exploited to measure the stimulusdriven activity. Here we describe stimuli in which vertical and horizontal disparities trade during the construction of percepts of slanted surfaces, yielding stimulus equivalence classes. Equivalence class membership changed after a change of vergence eye posture alone, without changes to the retinal images. A formal correspondence can be drawn between these ?perceptual metamers? and more familiar ?sensory metamers? such as color metamers. 1 Introduction Two types of perceptual process might, in principle, map physically different visual stimuli onto the same percept. First, the visual system has a host of constancy mechanisms that extract information about the visual environment across uninteresting changes in the proximal stimulus. Some of these mechanisms could be ?leak-proof,? leaving no trace of the original differences between the stimuli. Second, the visual system must combine information from redundant cues if it is to build percepts robustly. Recent cue conflict experiments have shown that the visual system?s estimate of a scene parameter, as evinced in a visual percept, is often simply a weighted average of the parameter as specified by each cue separately [1][2]. Thus, a properly balanced cue-conflict stimulus might come to evoke the same percept as a ?natural? or cue-concordant stimulus. * http://psych.upenn.edu/~backus Here, random-dot stereograms will be used to argue that leak-proof versions of both types of process exist. When a vertical magnifier is placed before one eye, a truly frontoparallel surface appears slanted. Adding horizontal magnification in the same eye restores frontoparallel appearance. The original stimulus and the magnified stimulus therefore have different patterns of binocular disparity but give rise to similar judgments of surface slant [3]. We show here that such stimuli are perceptually indistinguishable to practiced observers in a psychophysical discrimination task, which implies the loss of some disparity information. This loss could occur, first, in a well-studied constancy mechanism that uses vertical disparity to correct the depth relief pattern associated with horizontal disparity [4]. However, the amount of horizontal magnification needed to null vertical magnification is less than would be predicted from use of this constancy mechanism alone; a second constancy mechanism exists that corrects horizontal disparities by using felt eye position, not vertical disparity [5]. Adding vertical magnification without changing eye position therefore creates a cue conflict stimulus. We show here that the amount of horizontal magnification needed to null the vertical magnification changes with the vergence posture of the eyes, which implies that both types of process (constancy and cue combination) are leak-proof across certain ranges of variation (magnifications) in these stereoscopic stimuli. 2 Stereoscopic slant perception: review of theory The stereo component of the perceived slant of a random-dot surface can be modeled as the visual system?s weighted average of two stereo slant estimates [5][ 6]. Horizontal disparity is ambiguous because it depends not only on surface slant, but also on surface patch location relative to the head. One stereo estimator resolves this ambiguity using vertical disparity (images are vertically larger in the closer eye), and the other resolves it using felt eye position. Vertical magnification in one eye thus creates a cue-conflict because it affects only the estimator that uses vertical disparity. The two stereo estimators have different relative reliability at different distances, so the weights assigned to them by the visual system changes as a function of distance [7]. Since vergence eye posture is a cue to distance [8], one might predict that ?perceptually metameric? stereo stimuli, if they exist, will lose their metameric status after a pure change of vergence eye posture that preserves the metamers? retinal images [9]. We shall now briefly describe the two stereoscopic slant estimators. This theory is covered elsewhere in greater detail [5]. Although surface slant has two components (slant and tilt [10]), we will consider only slant about a vertical axis. The arguments can be extended to slant about axes of arbitrary orientation [5]. The visual signals used in stereoscopic slant perception can be conveniently parameterized by four numbers [5]. Each can be considered a signal. A surface patch typically gives rise to all four signals. Two signals are the horizontal gradient of horizontal disparity, and the vertical gradient of vertical disparity, which we parameterize as horizontal size ratio (HSR) and vertical size ratio (VSR), respectively, in the manner of Rogers and Bradshaw [11]. They are defined as the horizontal (or vertical) size of the patch in the left eye, divided by the horizontal (or vertical) size in the right eye. These two signals must be measured from the retinal images. The two remaining signals are the headcentric azimuth and vergence of the surface patch. These signals can be known either by measuring the eyes? version and vergence, respectively, or from the retinal images [12]. A very good approximation that relates surface slant to horizontal disparity and VSR is: S HSR,VSR = -tan-1 [ 1 ln HSR ? VSR ] Equation 1 where ? is the vergence of the surface patch in radians. We call this method of slant estimation slant from HSR and VSR. A very good approximation that relates surface slant to horizontal disparity and azimuth is: S HSR,EP = -tan-1 [ 1 ln HSR - tan? ? ] Equation 2 where ? is the azimuth of the surface patch. We call this method of slant estimation slant from HSR and eye position on the supposition that azimuth per se is known to the visual system primarily through measurement of the eyes? version. Each estimator uses three of the four signals available to estimate surface slant from horizontal disparity. Nonstereo slant estimates can be rendered irrelevant by the choice of task, in which case perceived slant is a weighted average of the slants predicted from these two stereoscopic slant estimates [5, 6]. In principle, the reliability of slant estimation by HSR and eye position is limited at short viewing distances (large ?) by error in the measurement of ?. Slant from HSR and VSR, on the other hand, continues to become more reliable as viewing distance decreases. If one assumes that the visual system knows how reliable each estimator is, one would predict that greater weight is given to the HSR and VSR estimate at near than at far distances, and this is in fact the case [7]. Whether each estimate is separately computed in its own neural process, and then given a weight, is not known. A maximum a posteriori Bayesian scheme that simply estimates the most likely slant given the observed signals behaves in a similar fashion as the weighted estimates model, though actual likelihood density (probability per deg of slant) is extraordinarily small in the case of stimuli that contain large cue conflicts [9]. The real visual system does not flinch, but instead produces a slant estimate that looks for all the world like a weighted average. It remains a possibility therefore that optimal slant estimation is implemented as a weighted combination of separate estimates. We have now developed the theory to explain why HSR and VSR trade with each other at the ?constancy? level of a single estimator (Equation 1), and why natural stimuli might appear the same as cue conflict stimuli (weighted averaging of estimates derived from exploitation of Equations 1 and 2, respectively). We next describe experiments that tested whether magnified (cue conflict) stimuli are distinguishable from natural (concordant) stimuli. 3 Existence of stereoscopic metamers Stimuli were sparse random dot stereograms (RDS) on a black background, 28 deg in diameter, presented directly in front of the head using a haploscope. Observers performed a forced choice task with stimuli that contained different amounts of unilateral vertical and horizontal magnification. Vertical magnification was zero for the ?A? stimuli, and 2% in the right eye for the ?B? stimuli (1% minification in the left eye and 1% magnification in the right eye). Horizontal magnification was set at the value that nulled apparent slant in ?A? stimuli (i.e. approximately 0%), and took on a range of values in ?B? stimuli. Each trial consisted of two ?A? stimuli and one ?B? stimulus. The observer?s task was to determine whether the three stimuli were presented in AAB or BAA order [13], i.e., whether the stimulus with vertical magnification was first or last of the three stimuli. Each stimulus was presented for 0.5 sec. Each stimulus was generated using a fresh set of 200 randomly positioned dots. Each dot had a circular raised cosine luminance profile that was 30 arcmin in diameter. Three observers participated, including the author. Results are shown in Figure 1. vMags = 0% and 2% N = 40 trials per datapoint Percent Correct 100 BTB MJN JRF 80 60 40 -3 -2 -1 0 -1.4 -1 -0.6 -2 0 2 Horiz mag in left eye in stimulus B (%) Figure 1. Observers are unable to distinguish 0% and 2% unilateral vertical magnification when unilateral horizontal magnification is added as well. Open squares show the horizontal magnification that evoked zero perceived slant under 2% vertical magnification. For each observer, there was a value of horizontal magnification that, when added to the ?B? stimulus, rendered it indistinguishable from the ?A? stimulus. This is shown in Figure 1 by the fact that performance drops to chance (50%) at some value of horizontal magnification. From this experiment it is evident that stimuli with very different disparity patterns can be made perceptually indistinguishable in a forcedchoice task with well-practiced observers. 3.1 Experimental conditions necessary for stereo metamers Several properties of the experiment were essential to the effect. First, the vertical magnification must not be to large. At large vertical magnifications it is still possible to null apparent slant, but the stimuli are distinguishable because the dots themselves look different (they look as though blurred in the vertical direction). Two out of three observers were able to distinguish the ?A? and ?B? stimuli 100% of the time when the vertical magnification was increased from 2% to 5%. Second, observers must be instructed to maintain fixation. If left and right saccades are allowed, the ?B? stimulus appears slanted in the direction predicted by its horizontal magnification. This is a rather striking effect?the surface appears to change slant simply because one starts looking about. This effect was not found previously [14] but is predicted as a consequence of sequential stereopsis [15]. Finally, if the stimuli are shown for more than about 1 sec it is possible to distinguish ?A? and ?B? stimuli by making vertical saccades from the top to the bottom of the stimulus, by taking advantage of the fact that in forward gaze, vertical saccades have equal amplitude in the two eyes [16]. For ?B? stimuli only, the dots are diplopic (seen in double vision) immediately after a saccade to the top (or bottom) of the stimulus. An automatic vertical vergence eye movement then brings the dots into register after about 0.5 sec. At that point a saccade to the bottom (or top) of the stimulus again causes diplopia. 4 Breaking metamerization though change of vergence eye posture In the haploscope it was possible to present unchanged retinal images across a range of vergence eye postures. Stimuli that were metameric to each other with the eyes verged at 100 cm were presented again with the eyes verged at 20 cm. For three out of four observers, the images were then distinguishable. Figure 2 illustrates this effect schematically, and Figure 3 quantifies it by plotting the amount of horizontal magnification that was needed to null apparent slant at each of the two vergence angles for one observer (left panel) and all four observers (right panel). HSR & VSR HSR & eye pos Percept Vmag Hmag Figure 2. Schematic illustration of the effect of distance in the slant-nulling task. First panel: both stereoscopic methods of estimating slant indicate that the surface is frontoparallel, and it appears so. Second panel: a vertical magnifier is placed before one eye, changing the estimate that uses vertical disparity, but not the estimate that uses eye position. The resulting percept is a weighted average of the two. Third panel: horizontal magnification is added until the surface appears frontoparallel again. At this point the two stereo estimates have opposite sign. Fourth panel: increasing the apparent distance to the stimulus (by decreasing the vergence) scales up both estimates by the same factor. The surface no longer appears frontoparallel because the weighting of the estimates has changed. Horiz magnification to null slant (%) Vertical magnification: ?2% 2.8 20 cm 100 cm 2.4 20 cm 2.0 1.6 1.2 100 cm 100 cm 0 5 10 15 20 25 30 35 40 Trial MJ N BT CS JRF B Subject Figure 3. When the eyes were verged at 100 or 20 cm distance, different amounts of horizontal magnification were needed to null the slant induced by vertical magnification. Left: 10 settings that nulled slant at 100 cm, followed by 20 settings at 20 cm, followed by 10 at 100 cm (observer BTB). Right: three out of four observers show an effect of vergence per se. Error bars are SEs of the mean. 5 Comparison of perceptual and sensory metamers The stimuli described here appear the same as a result of perceptual computations that occur well after transduction of light energy by the photoreceptors. Physically different stimuli that are transduced identically might be dubbed sensory metamers. One example of a sensory metamer is given by the trade between intensity and duration for briefly flashed lights (Bloch?s Law [17]): two flashes containing the same number of photons are indistinguishable if their durations are both less than 10 msec. Another example of sensory metamerization, that we will now consider in greater detail, is the traditional color metamer. The three cone photoreceptor types can support color vision because they are sensitive to different wavelengths of light. However, each cone type responds to a range of wavelengths, and two lights with different spectra may activate the three cone types identically. From that point on, the lights will be indistinguishable within the nervous system. (See [18] for a review of color metamers). Table 1 summarizes several properties of color metamers, and analogous properties of our new stereo metamers. We can approximate the visible spectrum of a light by sampling its power within N different wavelength intervals, where N is large. Thus light t can be represented by an Nx1 vector. Light t? is metameric to t if Bt? = Bt, where B is the 3xN matrix whose rows represent the spectral sensitivities of the three cone mechanisms [19]. The transformation that maps one stereo metamer to another is simply a scaling of one eyes? image in the vertical and horizontal directions, with less scaling typically needed in the horizontal than vertical direction. Let u and v represent the x and y disparity, respectively, so that [u v] is a function of location (x,y) within the cyclopean image. Then two random-dot image pairs (representing flat surfaces slanted about a vertical axis) will be metameric if their disparity patterns, [u? v?] and [u v], are related to each other by [u? v?] = [u(1+m) v(1+n)], where m and n are small (on the order of 0.01), with m/n equal to the weight of SHSR,VSR in the final slant estimate. Table 1: properties of color and stereo metamers PROPERTY COLOR METAMERS STEREO METAMERS. Metamer type: Sensory Perceptual Site of loss: Peripheral Central (two places) Loss process: Transduction Metameric class formation: Lights t? and t are metameric iff Bt? = Bt, where B is the 3xN matrix of cone spectral sensitivities Computation Disparity map [u v] is metameric to [u? v?] iff [u? v?] = [u(1+m) v(1+n)] where m and n are small and in the proper ratio Dimensionality reduction: N? 3 loss of 1 degree of freedom Etiology: Capacity limit Recovery of scene parameter Computation of surface slant removes one dimension from the set of all physical stimuli. Depending how the problem is framed, this is a reduction from 2 dimensions (HSR and VSR) to one (slant), or from many dimensions (all physical stimuli that represent slanted surfaces) to one fewer dimensions. While color and stereo metamers can be described as sensory and perceptual, respectively, the boundary between these categories is fuzzy, as is the boundary between sensation and perception. Would motion metamers based on ?early? motion detectors be sensory or perceptual? What of stimuli that look identical to retinal ganglion cells, after evoking different patterns of photoreceptor activity? While there is a real distinction to be made between sensory and perceptual metamers, but not all metamers need be easily categorized as one or the other. 5.1 The metamer hierarchy Loftus [20] makes a distinction reminiscent of the one made here, between ?memory metamers? and ?perceptual metamers,? with memory metamers being stimuli that evoke distinguishable percepts during live viewing, but that become indistinguishable after mnemonic encoding. Thus, Loftus classified as ?perceptual? both our perceptual and sensory metamers. Figure 4 suggests how the three concepts are related. In this framework, color and stereo metamers are both perceptual metamers, but only color metamers are sensory metamers. Memory metamers Perceptual metamers (e.g. stereo) Sensory metamers (e.g. color) Figure 5. The metamer hierarchy. 6 Conclusions At each vergence eye posture it was possible to create stereoscopic stimuli with distinct disparity patterns that were nonetheless indistinguishable in a forced choice task. Stimuli that were metamers with the eyes in one position became distinguishable after a change of vergence eye posture alone, without changes to the retinal images. We can conclude that horizontal disparity per se is lost to the visual system after combination with the other signals that are used to interpret it as depth. Presumably, stereo metamers have distinguishable representations in primary visual cortex?one suspects this would be evident in evoked potentials or fMRI. The loss of information that renders these stimuli metameric probably occurs in two places. First, there appears to be a leak-proof ?constancy? computation in which vertical disparity is used to correct horizontal disparity (Equation 1). The output of this computation is unaffected if equal amounts of horizontal and vertical magnification are added to one eyes? image. However, the estimator that uses felt eye position can distinguish these stimuli, because their horizontal size ratios differ. Thus a second leak-proof step must occur, in which slant estimates are combined in a weighted average. It seems reasonable to call these stimuli ?perceptual metamers,? by analogy with, and to distinguish them from, the traditional ?sensory? metamerization of colored lights. Acknowledgments This work was supported by startup funds provided to the author by the University of Pennsylvania. The author thanks Mark Nolt for help conducting the experiments, Rufus Frazer for serving as an observer, and Jack Nachmias and David Brainard for comments on an earlier draft of this paper. References 1. Clark, J.J. and A.L. Yuille, Data fusion for sensory information processing systems. 1990, Boston: Kluwer. 2. Landy, M.S., et al., Measurement and modeling of depth cue combination: in defense of weak fusion. Vision Research, 1995. 35(3): p. 389-412. 3. Ogle, K.N., Induced size effect. I. A new phenomenon in binocular space perception associated with the relative sizes of the images of the two eyes. Archives of Ophthalmology, 1938. 20: p. 604-623. 4. G?rding, J., et al., Stereopsis, vertical disparity and relief transformations. Vision Res, 1995. 35(5): p. 703-22. 5. Backus, B.T., et al., Horizontal and vertical disparity, eye position, and stereoscopic slant perception. Vision Res, 1999. 39(6): p. 1143-70. 6. Banks, M.S. and B.T. Backus, Extra-retinal and perspective cues cause the small range of the induced effect. Vision Res, 1998. 38(2): p. 187-94. 7. Backus, B.T. and M.S. Banks, Estimator reliability and distance scaling in stereoscopic slant perception. Perception, 1999. 28(2): p. 217-42. 8. Foley, J.M., Binocular distance perception. Psychol Rev, 1980. 87(5): p. 411-34. 9. Backus, B.T. and M.J. Nolt, Analysis of stereoscopic metamers. Journal of Vision (Vision Sciences conference supplement), 2001. 1: p. in press. 10. Stevens, K.A., Slant-tilt: the visual encoding of surface orientation. Biol Cybern, 1983. 46(3): p. 183-95. 11. Rogers, B.J. and M.F. Bradshaw, Vertical disparities, differential perspective and binocular stereopsis. Nature, 1993. 361(6409): p. 253-5. 12. Mayhew, J.E. and H. Longuet-Higgins, C, A computational model of binocular depth perception. Nature, 1982. 297(5865): p. 376-378. 13. Calkins, D.J., J.E. Thornton, and E.N. Pugh, Jr., Monochromatism determined at a long-wavelength/middle-wavelength cone- antagonistic locus. Vision Res, 1992. 32(12): p. 2349-67. 14. van Ee, R. and C.J. Erkelens, Temporal aspects of binocular slant perception. Vision Res, 1996. 36(1): p. 43-51. 15. Enright, J.T., Sequential stereopsis: a simple demonstration. Vision Res, 1996. 36(2): p. 307-12. 16. Schor, C.M., J. Gleason, and D. Horner, Selective nonconjugate binocular adaptation of vertical saccades and pursuits. Vision Res, 1990. 30(11): p. 1827-44. 17. Barlow, H.B., Temporal and spatial summation in human vision at different backgound intensities. Journal of Physiology, 1958. 141: p. 337-350. 18. Wandell, B.A., Foundations of vision. 1995, Sunderland, MA: Sinauer Associates. 19. Baylor, D.A., B.J. Nunn, and J.L. Schnapf, Spectral sensitivity of cones of the monkey Macaca fascicularis. J Physiol, 1987. 390: p. 145-60. 20. Loftus, G.R. and E. Ruthruff, A theory of visual information acquisition and visual memory with special application to intensity-duration trade-offs. J Exp Psychol Hum Percept Perform, 1994. 20(1): p. 33-49.
2067 |@word trial:3 exploitation:1 briefly:2 version:3 middle:1 seems:1 open:1 reduction:2 disparity:27 mag:1 practiced:2 past:1 must:5 slanted:5 reminiscent:1 physiol:1 visible:1 remove:1 drop:1 fund:1 discrimination:1 alone:3 cue:15 fewer:1 nervous:1 short:1 metamerization:3 colored:1 draft:1 location:2 become:2 diplopia:1 differential:1 fixation:1 combine:1 behavioral:1 manner:1 rding:1 upenn:2 themselves:1 brain:1 decreasing:1 resolve:2 actual:1 increasing:1 provided:1 estimating:1 panel:6 transduced:1 null:6 what:1 cm:11 psych:2 monkey:1 fuzzy:1 developed:1 magnified:2 dubbed:1 transformation:2 temporal:2 appear:2 evoking:1 interestingness:1 before:2 vertically:1 limit:1 consequence:1 encoding:2 approximately:1 might:5 black:1 studied:1 equivalence:2 evoked:2 suggests:1 limited:1 range:5 jrf:2 forcedchoice:1 acknowledgment:1 lost:1 area:1 physiology:1 nulled:2 suggest:1 cannot:1 onto:1 pugh:1 live:1 cybern:1 map:3 duration:3 recovery:1 immediately:1 pure:1 estimator:9 aab:1 higgins:1 variation:1 analogous:1 antagonistic:1 construction:1 tan:3 hierarchy:2 us:6 pa:1 associate:1 magnification:30 continues:1 ep:1 constancy:7 observed:1 bottom:3 parameterize:1 trade:4 backus:7 decrease:1 movement:1 balanced:1 benjamin:1 environment:1 leak:5 stereograms:2 yuille:1 creates:2 po:1 easily:1 represented:1 forced:2 distinct:1 describe:3 activate:1 startup:1 rds:1 formation:1 extraordinarily:1 apparent:4 whose:1 larger:1 s:1 final:1 thornton:1 advantage:1 took:1 adaptation:1 iff:2 verged:3 macaca:1 arcmin:1 double:1 produce:1 help:1 depending:1 brainard:1 measured:2 mayhew:1 implemented:1 predicted:4 c:1 come:1 implies:2 indicate:1 frontoparallel:5 direction:4 sensation:1 differ:1 stevens:1 correct:3 human:1 viewing:3 rogers:2 summation:1 considered:1 exp:1 presumably:1 predict:2 early:1 perceived:3 estimation:4 lose:1 sensitive:1 create:1 weighted:9 offs:1 rather:1 schnapf:1 ax:1 derived:1 properly:1 likelihood:1 posteriori:1 membership:1 typically:2 bt:5 sunderland:1 selective:1 orientation:2 restores:1 raised:1 spatial:1 special:1 equal:3 sampling:1 identical:1 look:4 survive:1 fmri:1 stimulus:63 primarily:1 mjn:1 randomly:1 preserve:1 familiar:1 relief:2 maintain:1 freedom:1 possibility:2 vsr:11 circular:1 truly:1 yielding:1 light:10 bloch:1 closer:1 necessary:1 re:7 increased:1 earlier:1 modeling:1 measuring:1 uninteresting:1 azimuth:4 front:1 proximal:1 combined:1 thanks:1 density:1 sensitivity:3 corrects:1 gaze:1 again:3 ambiguity:1 central:1 containing:1 horiz:2 potential:1 photon:1 retinal:8 sec:3 blurred:1 register:1 depends:1 performed:1 observer:15 start:1 square:1 became:1 conducting:1 percept:11 judgment:1 rufus:1 weak:1 bayesian:1 none:1 unaffected:1 classified:1 explain:1 datapoint:1 detector:1 energy:1 nonetheless:1 acquisition:1 proof:5 associated:2 radian:1 gleason:1 color:11 dimensionality:1 amplitude:1 positioned:1 appears:7 nonconjugate:1 though:3 binocular:7 until:1 hand:1 horizontal:32 brings:1 effect:8 concept:1 consisted:1 barlow:1 contain:1 assigned:1 flashed:1 indistinguishable:7 during:2 ambiguous:1 cosine:1 btb:2 evident:2 motion:2 percent:1 image:13 jack:1 wandell:1 behaves:1 physical:2 tilt:2 kluwer:1 interpret:1 measurement:3 slant:46 framed:1 automatic:1 had:1 dot:9 reliability:3 longer:1 surface:22 cortex:1 own:1 recent:1 perspective:2 irrelevant:1 driven:1 certain:1 exploited:1 seen:1 greater:3 fascicularis:1 determine:1 redundant:1 signal:10 relates:2 mnemonic:1 long:1 divided:1 host:1 schematic:1 baa:1 vision:15 physically:2 represent:3 cell:1 background:1 schematically:1 separately:2 participated:1 interval:1 leaving:1 extra:1 archive:1 probably:1 comment:1 subject:1 induced:3 suspect:1 call:3 ee:1 near:1 identically:2 affect:1 psychology:1 pennsylvania:2 opposite:1 whether:4 defense:1 etiology:1 stereo:15 render:1 unilateral:3 cause:2 covered:1 se:3 amount:6 schor:1 bradshaw:2 category:1 diameter:2 http:1 exist:3 stereoscopic:12 sign:1 per:5 serving:1 cyclopean:1 metameric:9 shall:1 four:6 drawn:1 loftus:3 changing:4 luminance:1 frazer:1 cone:7 angle:1 parameterized:1 fourth:1 striking:1 place:2 reasonable:1 patch:6 summarizes:1 scaling:3 followed:2 distinguish:6 correspondence:1 evinced:1 activity:4 occur:3 scene:2 flat:1 felt:3 aspect:1 argument:1 rendered:2 department:1 peripheral:1 combination:5 nachmias:1 jr:1 across:3 ophthalmology:1 rev:1 making:1 ln:2 equation:5 remains:1 previously:1 mechanism:6 needed:5 know:2 locus:1 available:1 pursuit:1 spectral:3 distinguished:1 robustly:1 existence:1 original:2 top:4 remaining:1 assumes:1 landy:1 build:2 unchanged:1 psychophysical:1 added:4 posture:8 occurs:1 hum:1 primary:1 traditional:2 responds:1 gradient:2 distance:11 separate:1 unable:1 capacity:1 argue:1 reason:1 fresh:1 modeled:1 illustration:1 ratio:4 demonstration:1 baylor:1 difficult:1 trace:1 rise:2 proper:1 perform:1 vertical:41 extended:1 looking:1 head:2 arbitrary:1 intensity:3 david:1 pair:1 specified:1 conflict:7 distinction:2 horner:1 able:1 bar:1 pattern:7 perception:11 reliable:2 including:1 memory:4 power:1 natural:3 representing:1 scheme:1 eye:39 axis:2 psychol:2 extract:1 philadelphia:1 foley:1 review:2 relative:3 law:1 sinauer:1 loss:6 interesting:1 analogy:1 clark:1 foundation:1 degree:1 principle:2 plotting:1 bank:2 row:1 elsewhere:1 changed:2 placed:2 last:1 supported:1 formal:1 taking:1 sparse:1 van:1 boundary:2 depth:4 xn:2 world:1 dimension:4 sensory:15 author:3 made:3 instructed:1 forward:1 far:1 approximate:1 status:1 evoke:3 deg:2 photoreceptors:1 conclude:1 stereopsis:4 spectrum:2 quantifies:1 vergence:15 why:2 table:2 mj:1 nature:2 longuet:1 constructing:1 profile:1 allowed:1 categorized:1 site:1 fashion:1 transduction:2 position:9 msec:1 hsr:14 perceptual:15 breaking:1 third:1 weighting:1 down:1 supposition:1 fusion:2 exists:1 essential:1 adding:2 sequential:2 supplement:1 metamers:34 perceptually:3 illustrates:1 boston:1 distinguishable:6 simply:4 appearance:1 likely:1 wavelength:5 ganglion:1 visual:18 conveniently:1 contained:1 saccade:6 chance:1 ma:1 flash:1 change:10 determined:1 averaging:1 experimental:1 concordant:2 photoreceptor:2 nx1:1 support:1 mark:1 phenomenon:1 tested:1 biol:1
1,171
2,068
Learning hierarchical structures with Linear Relational Embedding Alberto Paccanaro Geoffrey E. Hinton Gatsby Computational Neuroscience Unit UCL, 17 Queen Square, London, UK alberto,hinton  @gatsby.ucl.ac.uk Abstract We present Linear Relational Embedding (LRE), a new method of learning a distributed representation of concepts from data consisting of instances of relations between given concepts. Its final goal is to be able to generalize, i.e. infer new instances of these relations among the concepts. On a task involving family relationships we show that LRE can generalize better than any previously published method. We then show how LRE can be used effectively to find compact distributed representations for variable-sized recursive data structures, such as trees and lists. 1 Linear Relational Embedding Our aim is to take a large set of facts about a domain expressed as tuples of arbitrary symbols in a simple and rigid syntactic format and to be able to infer other ?common-sense? facts without having any prior knowledge about the domain. Let us imagine a situation in which we have a set of concepts and a set of relations among these concepts, and that our data consists of few instances of these relations that hold among the concepts. We want to be able to infer other instances of these relations. For example, if the concepts are the people in a certain family, the relations are kinship relations, and we are given the facts ?Alberto has-father Pietro? and ?Pietro has-brother Giovanni?, we would like to be able to infer ?Alberto has-uncle Giovanni?. Our approach is to learn appropriate distributed representations of the entities in the data, and then exploit the generalization properties of the distributed representations [2] to make the inferences. In this paper we present a method, which we have called Linear Relational Embedding (LRE), which learns a distributed representation for the concepts by embedding them in a space where the relations between concepts are linear transformations of their distributed representations. Let us consider the case in which all the relations are binary, i.e. involve two concepts. In this case our data consists of triplets          , and the problem we are trying to solve is to infer missing triplets when we are given only few of them. Inferring a triplet is equivalent to being able to complete it, that is to come up with one of its elements, given the other two. Here we shall always try to complete the third element of the triplets 1 . LRE will then represent each concept in the data as a learned vector in a 1 Methods analogous to the ones presented here that can be used to complete any element of a triplet can be found in [4]. Euclidean space and each relationship between the two concepts as a learned matrix that maps the first concept into an approximation to the second concept. Let us assume that our data consists of such triplets containing distinct concepts and binary relations.   will denote the set of  -dimensional We shall call this set of triplets ;   vectors corresponding to the concepts, and    the set of   matrices corresponding to the relations. Often we shall need to indicate the vectors and the matrix which correspond to the concepts and the relation in a certain triplet  . In this case we shall denote the vector corresponding to the first concept with , the vector corresponding to the second concept with and the matrix corresponding to the relation with . We shall therefore write the triplet  as    where  and . The operation that relates a pair   to a vector is the matrix-vector multiplication, , which produces an approximation to . If for every triplet   we think of  as a noisy version of one of the concept vectors, then one way to learn an embedding is to maximize the probability that it is a noisy version of the correct completion, . We imagine that a concept has an average location in the space, but that each ?observation? of the concept is a noisy realization of this average location. Assuming spherical Gaussian noise with a variance of on each dimension, the discriminative goodness function that corresponds to the log probability of getting the right completion, summed over all training triplets is:             !"    %$     $ &        #         !      ' ()+* ,  . 1 ( 24365 . 6798:8 ; <>= ?<@7BAC<@8:E 8 D F (1) 0/  E+F0GIH 798:8 ; <>= ?<J7 8:8 D 1 where is the number of triplets in  having the first two terms equal to the ones of , but  differing in the third term . , Learning based on maximizing with respect to all the vector and matrix components has given good results, and has proved successful , in generalization as well [5]. However, when we learn an embedding by maximizing , we are not making use of exactly the information that we have in the triplets. For each triplet , we are making the vector representing the   $ B , while correct completion K  more probable than any other concept vector given , the triplet states that  $ & must be equal to L . The numerator of does exactly this, but we also have the denominator, which is necessary in order to stay away from the trivial M solution . We noticed however that the denominator is critical at the beginning of the learning, but- as the vectors and matrices differentiate we could gradually lift this burden,   $ &QPRS O to become the real goal of the learning. To do this we allowing N  O 0  / modify the discriminative function to include a parameter T , which is annealed from ( to U V during learning :  . 1 ( 24365 Y . 679W079; W0<J; = ?<J<J= ?7B<>A%7 <XE W0F D W0DX[]\ (2)  0/ E+FZGIH We would like our system to assign equal probability to each of the correct completions. de4f eBk Thelm l tokQapproximate discrete probability distribution that we want is therefore: ^_a`cd b N b6g6hji the w@ykzl"u@{Xm in n . Our system implements where is the discrete delta function and rangesu>vx qJsIt h over the vectorswhere b discreteg probability distribution: o _ ` pr is the normalization factor. | The }X~ ? factor in eq.1 ensures that we are minimizing the Kullback-Leibler divergence between ^ and o . v ? w y ? approach The obvious an embedding would be to minimize the sum of squared distances ? overto find between and all the triplets, with respect to all the vector and matrix components. i Unfortunately this minimization (almost) always causes all of the vectors and matrices to collapse to the trivial ? solution. one-to-many relations we must not decrease the value of ? all the way to ? , because this v ? wJFor y ? cause would some concept vectors to become coincident. This is because the only way to make equal to  ? different vectors, is by collapsing them onto a unique vector.   2  3   4  2 3 4 V V During learning this function (for Goodness) is maximized with respect to all the vector and matrix components. This gives a much better generalization performance than the one obtained by just maximizing . The results presented in the next sections were obtained by maximizing using gradient ascent. All the vector and matrix components were updated simultaneously at each iteration. One effective method of performing the optimization is conjugate gradient. Learning was fast, usually requiring only a few hundred updates. It is worth pointing out that, in general, different initial configurations and optimization algorithms caused the system to arrive at different solutions, but these solutions were almost always very similar in terms of generalization performance. , 2 LRE results Here we present the results obtained applying LRE to the Family Tree Problem [1]. In this problem, the data consists of people and relations among people belonging to two families, one Italian and one English, shown in fig.1 (left) 5 . All the information in these trees can be represented in simple propositions of the form          . Using the relations father, mother, husband, wife, son, daughter, uncle, aunt, brother, sister, nephew, niece there are 112 such triplets in the two trees. Fig.1 (right) shows the embedding obtained after training with LRE. Notice how the Italians are linearly separable from the English people. From the Hinton diagram, we can see that each member of a family is symmetric to the corresponding member in the other family. The sign of the third component of the vectors is (almost) a feature for the nationality. When testing the generalization 1 Margaret = Arthur 9 8 2 7 Christopher = Penelope Andrew = Christine 10 Victoria = James 4 3 Colin 6 13 Charlotte 19 Jennifer = Charles 2 0 12 14 Aurelio = Maria 5 11 ?2 20 ?5 Bortolo = Emma 0 ?5 0 5 English Italians 5 Grazia = Pierino 21 22 15 18 Giannina = Pietro Alberto Doralice = Marcello 16 23 17 Mariemma 24 Figure 1: Left: Two isomorphic family trees. The symbol ?=? means ?married to?. Right Top: layout of the vectors representing the people obtained for the Family Tree Problem in 3D. Vectors end-points are indicated by *, the ones in the same family tree are connected to each other. All triplets were used for training. Right Bottom: Hinton diagrams of the 3D vectors shown above. The vector of each person is a column, ordered according to the numbering on the tree diagram on the left. (6(*    $ &   performance, for each triplet in the test set    , we chose as completion the concepts  according to their probability, given . The system was generally able to complete correctly all triplets even when  of them, picked at random, had been left out during training. These results on the Family Tree Problem are much better than the ones obtained using any other method on the same problem: Quinlan?s FOIL [7] could generalize on  triplets, while Hinton (1986) and O?Reilly (1996) made one or more errors when only  test cases were held out during training. (6(* * 5 The names of the Italian family have been altered from those originally used in Hinton (1986) to match those of one of the author?s family. For most problems there exist triplets which cannot be completed. This is the case, for example, of (Christopher, father, ?) in the Family Tree Problem. Therefore, here we argue that it is not sufficient to test generalization by merely testing the completion of those complete-able triplets which have not been used for training. The proper  test for generalization is to see how the system completes any triplet of the kind    where ranges over the concepts and R over the relations. We cannot assume to have knowledge of which triplets admit a completion, and which do not. To our knowledge this issue has never been analyzed before (even though FOIL handles this problem correctly). To do this the system needs a way to indicate when a triplet does not admit a completion. Therefore, once the maximization of is terminated, we build a new probabilistic model around the solution which has been found. This new model is constituted, for each relation, of a mixture of identical spherical Gaussians, each centered on a concept vector, and a Uniform distribution. The Uniform distribution will take care of the ?don?t know? answers, and will be competing with all the other Gaussians, each representing a concept vector. For each relation the Gaussians have different variances and the Uniform a different height. The parameters of this probabilistic model are, for each relation , the variances of the Gaussians and the relative density under the Uniform distribution, which we shall write    as    . These parameters are learned using a validation set, which will be the union of a set of complete-able (positive) triplets  and a set of pairs which  cannot be and completed  (negative); that is            where  indicates the fact that the result of applying relation  to  does not belong  to . This is done by maximizing the following discriminative goodness function  over the validation set :    V   ZP ; ;I) * ;    > /   c   Z/     P   D D  . 2 4 6 3 5   .  ZP O  $   P   O   P   D   / * ; D E+F0GIH  P W0;  = ? . 1 ( $ 2 3I5  7B  AIW D D  $ " P   (3)  .  O O D       Z P !      Z P /   * ; D E+F GIH with respect to the  > else  is kept fixed. Having ; and ; parameters, while everything we compute the learned these parameters, in order to complete any triplet  probability distribution over each of the  Gaussians and the Uniform distribution given $  . The system then chooses a vector or the ?don?t know? answer according to those probabili                     ties, as the completion to the triplet. (* We used this method on the Family Tree Problem using a train, test and validation sets built in the following way. The test set contained positive triplets chosen at random, but such that there was a triplet per relation. The validation set contained a group of positive and a group of negative triplets, chosen at random and such that each group had a triplet per relation. The train set contained the remaining  positive triplets. After learning a distributed representation for the entities in the data by maximizing over the training set, we learned the parameters of the probabilistic model by maximizing  over the validation set. The resulting system was able to correctly complete all the  possible triplets    . Figure 2 shows the distribution of the probabilities when completing one complete-able and one uncomplete-able triplet in the test set. (* V (* *  > LRE seems to scale up well to problems of bigger size. We have used it on a much bigger version of the Family Tree Problem, where the family tree is a branch of the real family  # people over $ generations. Using the same set tree of one of the authors containing " of relations used in the Family Tree Problem, there is a total of %  positive triplets. After learning using a training set of $  positive triplets, and a validation set constituted (* I* Charlotte uncle Emma aunt 1 1 0.8 0.8 0.6 0.6 0.4 0.4 0.2 0.2 0 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 Figure 2: Distribution of the probabilities assigned to each concept for one complete-able (left) and one uncomplete-able (right) triplet written above each diagram. The completeable triplet has two correct completions but neither of the triplets had been used for training. Black bars from to  are the probabilities of the people ordered according to the numbering in fig.1. The last grey bar on the right, is the probability of the ?don?t know? answer. ( * IU IU by positive and negative triplets, the system is able to complete correctly almost all the possible triplets. When many completions are correct, a high probability is always assigned to each one of them. Only in few cases is a non-negligible probability assigned to some wrong completions. Almost all the generalization errors are of a specific form. The system appears to believe that ?brother/sister of? means ?son/daughter of parents of?. It fails to model the extra restriction that people cannot be their own brother/sister. On the other hand, nothing in the data specifies this restriction. 3 Using LRE to represent recursive data structures In this section, we shall show how LRE can be used effectively to find compact distributed representations for variable-sized recursive data structures, such as trees and lists. Here we discuss binary trees, but the same reasoning applies to trees of any valence. The approach is inspired by Pollack?s RAAM architecture [6]. A RAAM is an auto-encoder which is trained using backpropagation. Figure 3 shows the architecture of the network for binary trees. The system can be thought as being composed of two networks. The first one, called ~ l ~r Reconstructor ~ l ~r R1 R1 R2 R2 C1 C2 Verb Phrase Noun Phrase R2 C Compressor l r l R2 R1 r C1 Adjective a R1 C2 C1 Noun b Verb c C2 Noun d Figure 3: Left: the architecture of a RAAM for binary trees. The layers are fully connected. Adapted from [6]. Center: how LRE can be used to learn a representation for binary trees in a RAAM-like fashion. Right: the binary tree structure of the sentences used in the experiment. compressor encodes two fixed-width patterns into a single pattern of the same size. The second one, called reconstructor, decodes a compressed pattern into facsimiles of its parts, and determines when the parts should be further decoded. To encode a tree the network must learn as many auto-associations as the total number of non-terminal nodes in the tree. The codes for the terminal nodes are supplied, and the network learns suitable codes for the other nodes. The decoding procedure must decide whether a decoded vector represents a terminal node or an internal node which should be further decoded. This is done by using binary codes for the terminal symbols, and then fixing a threshold which is used for checking for ?binary-ness? during decoding. The RAAM approach can be cast as an LRE problem, in which concepts are trees, subtrees or leaves, or pairs of trees, sub-trees or leaves, and there exist relationships: implementing the compressor, and  and  which jointly implement the reconstructor (see fig.3). We can then learn a representation for all the trees, and the matrices by maximizing in eq.2. This formulation, which we have called Hierarchical LRE (HLRE), solves two problems encountered in RAAMs. First, one does not need to supply codes for the leaves of the trees, since LRE will learn an appropriate distributed representation for them. Secondly, one can also learn from the data when to stop the decoding process. In fact, the problem of recognizing whether a node needs to be further decoded, is similar to the problem of recognizing that a certain triplet does not admit a completion, that we solved in the previous section. While before we built an outlier model for the ?don?t know? answers, now we shall build one for the non-terminal nodes. This can be done by learning appropriate values of and  for relations  and  maximizing  in eq.3. The set of triplets    where is not a leaf of the tree, will play the role of the  set which appears in eq.3.  V !B        We have applied this method to the problem of encoding binary trees which correspond to sentences of  words from a small vocabulary. Sentences had a fixed structure: a noun phrase, constituted of an adjective and a noun, followed by a verb phrase, made of a verb and a noun (see fig.3). Thus each sentence had a fixed grammatical structure, to which we added some extra semantic structure in the following way. Words of each grammatical category were divided into two disjoint sets. Nouns were in girl, woman, scientist  or in  dog, doctor, lawyer  ; adjectives were in  pretty, young  or in  ugly, old  ; verbs were in  help, love  or in  hurt, annoy  . Our training set was consentences of the type:            stituted by  and of the type                , where the suffix indicates the set to which  each word type belongs. In this way, sentences of the kind ?pretty girl annoy scientist? were not allowed in the training set, and there were  possible sentences that satisfied the constraints which were implicit in the training set. We used HLRE to learn a distributed representation for all the nodes in the trees, maximizing using the sentences in the training set. In 7D, after having built the outlier model for the non-terminal symbols, given any root or internal node the system would reconstruct its children, and if they were non-terminal symbols would further decode each of them. The decoding process would always halt providing the correct reconstruction for all the sentences in the training set. The top row of fig.4 shows the distributed representations found for each word in the vocabulary. Notice how the and  sets of adjectives and verbs are almost symmetric with respect to the origin; the difference between the and  sets is less evident for the nouns, due to the fact that while there exists a restriction on which nouns can be used in position , there is no restriction on the nouns appearing in position  in the training sentences (see fig.3, right). We tested how well this system could generalize beyond the training set using the same procedure used by Pollack to enumerate the set of trees that RAAMs are able to represent [6]: for every pair of patterns for trees, first we encoded them into a pattern for a new higher level tree, and then we decoded this tree back into the patterns of the two sub-trees. If the norm of the difference between the original , then the tree and the reconstructed sub-trees was within a tolerance, which we set to could be considered to be well formed. The system shows impressive generalization performance: after training using the sentences, the four-word sentences it generates are all the  well formed sentences, and only those. It does not generate sentences which are either grammatically wrong, like ?dog old girl annoy?, nor sentences which violate semantic constraints, like ?pretty girl annoy scientist?. This is striking when compared to the poor generalization performance obtained by the RAAM on similar problems. As recognized by   V \  (U \   \ \  \ \ \ (  ( U \ *IU *+U ( T *+U T U ( Nouns Adjectives Verbs C1 ? C1 ? girl C1 ? C2 ? girl C2 ? C1 ? girl C2 ? C2 ? girl R1 ? R1 ? C1 ? C2 ? Nouns R1 ? R1 ? C1 ? C1 ? Adjectives R1 ? R1 ? C2 ? C2 ? Nouns R1 ? R1 ? C2 ? C1 ? Verbs Figure 4: For Hinton diagrams with multiple rows, each row relates to a word, in the following order - Adjectives: 1=pretty; 2=young; 3=ugly; 4=old  ; Nouns: 1=girl; 2=woman; 3=scientist; 4=dog; 5=doctor; 6=lawyer  ; Verbs: 1=help; 2=love; 3=hurt; 4=annoy  ; . ,  (higher), from  , ,  (lower). Top row: The disBlack bars separate  , tributed representation of the words in the sentences found after learning. Center row: The different contributions given to the root of the tree by the word ?girl? when placed in position  , ,  and  in the tree. Bottom row: The contribution of each leaf to the reconstruction of  , when adjectives, nouns, verbs and nouns are applied in positions  , ,  and  respectively. \  \ \  Pollack [6], this was almost certainly due to the fact that for the RAAMs the representation for the leaves was too similar, a problem that the HLRE formulation solves, since it learns their distributed representations. Let us try to explain why HLRE can generalize so well. The matrix can be decomposed into two sub-matrices,  and  , such that for any two children of a given node,  and  , we have:       , where ?;? denotes the concatenation operator. Therefore we have a pair of matrices, either     or     , associated to each link in the graph. Once the system has learned an embedding, finding a distributed representation for  a given tree amounts to multiplying the representation of its leaves by all the matrices found on all the paths from the leaves to the root, and adding them up. Luckily matrix multiplication is non-commutative, and therefore every sequence of words on its leaves can generate a different representation at the root node. The second row of fig.4 makes this point clear showing the different contributions given to the root of the tree by the word ?girl? , depending on its position in the sentence. A tree can be ?unrolled? from the root to  its leaves by multiplying its distributed representation using the matrices. We can now analyze how a particular leaf is reconstructed. Leaf  , for example, is reconstructed as: $Y [ $ R $          $ $        $ $ $   $ $ $            $ $     The third row of fig.4 shows the contribution of each leaf to the reconstruction of  , when adjectives, nouns, verbs and nouns are placed on leaves  , ,  and  respectively. We can see that the contributions from the adjectives, match very closely their actual distributed representations, while the contributions from the nouns in position  are negligible. This means that any adjective placed on  will tend to be reconstructed correctly, and that its reconstruction is independent of the noun we have in position  . On the other hand, the contributions from nouns and verbs in positions and  are non-negligible, and notice how those given by words belonging to the subsets are almost symmetric to those given by words in the  subsets. In this way the system is able to enforce the semantic agreement between words in positions  , and  . Finally, the reconstruction of  , when adjectives, nouns, verbs and nouns are not placed on leaves  , ,  and  respectively, assigns a very low probability to any word, and thus the system does not generate sentences which are not well formed. T 4 Conclusions Linear Relational Embedding is a new method for learning distributed representations of concepts and relations from data consisting of instances of relations between given concepts. It finds a mapping from the concepts into a feature-space by imposing the constraint that relations in this feature-space are modeled by linear operations. LRE shows excellent generalization performance. The results on the Family Tree Problem are far better than those obtained by any previously published method. Results on other problems are similar. Moreover we have shown elsewhere [4] that, after learning a distributed representation for a set of concepts and relations, LRE can easily modify these representations to incorporate new concepts and relations and that it is possible extract logical rules from the solution and to couple LRE with FOIL [7]. Learning is fast and LRE rarely converges to solutions with poor generalization. We began introducing LRE for binary relations, and then we saw how these ideas can be easily extended to higher arity relation by simply concatenating concept vectors and using rectangular matrices for the relations. The compressor relation for binary trees is a ternary relation; for trees of higher valence the compressor relation will have higher arity. We have seen how HLRE can be used to find distributed representations for hierarchical structures, and its generalization performance is much better than the one obtained using RAAMs on similar problems. It is easy to prove that, when all the relations are binary, given a sufficient number of dimensions, there always exists an LRE-type of solution that satisfies any set of triplets [4]. However, due to its linearity, LRE cannot represent some relations of arity greater than . This limitation can be overcome by adding an extra layer of non-linear units for representing the relations. This new method, called Non-Linear Relational Embedding (NLRE) [4], can represent any relation and has given good generalization results. * References [1] Geoffrey E. Hinton. Learning distributed representations of concepts. In Proceedings of the Eighth Annual Conference of the Cognitive Science Society, pages 1?12. Erlbaum, NJ, 1986. [2] Geoffrey E. Hinton, James L. McClelland, and David E. Rumelhart. Distributed representations. In David E. Rumelhart, James L. McClelland, and the PDP research Group, editors, Parallel Distributed Processing, volume 1, pages 77?109. The MIT Press, 1986. [3] Randall C. O?Reilly. The LEABRA model of neural interactions and learning in the neocortex. PhD thesis, Department of Psychology, Carnegie Mellon University, 1996. [4] Alberto Paccanaro. Learning Distributed Representations of Relational Data using Linear Relational Embedding. PhD thesis, Computer Science Department, University of Toronto, 2002. [5] Alberto Paccanaro and Geoffrey E. Hinton. Learning distributed representations by mapping concepts and relations into a linear space. In Pat Langley, editor, Proceedings of ICML2000, pages 711?718. Morgan Kaufmann, Stanford University, 2000. [6] Jordan B. Pollack. Recursive distributed representations. Artificial Intelligence, 46:77?105, 1990. [7] J. R. Quinlan. Learning logical definitions from relations. Machine Learning, 5:239?266, 1990.
2068 |@word niece:1 version:3 seems:1 norm:1 grey:1 initial:1 configuration:1 must:4 written:1 update:1 intelligence:1 leaf:15 beginning:1 node:11 location:2 toronto:1 lawyer:2 penelope:1 height:1 c2:11 become:2 supply:1 consists:4 prove:1 emma:2 love:2 annoy:5 nor:1 terminal:7 inspired:1 spherical:2 decomposed:1 actual:1 moreover:1 linearity:1 kinship:1 kind:2 differing:1 finding:1 transformation:1 nj:1 every:3 tie:1 exactly:2 wrong:2 uk:2 qpr:1 unit:2 before:2 positive:7 negligible:3 scientist:4 modify:2 encoding:1 tributed:1 path:1 black:1 chose:1 married:1 raams:4 collapse:1 range:1 unique:1 thel:1 testing:2 ternary:1 recursive:4 union:1 implement:2 backpropagation:1 procedure:2 langley:1 thought:1 reilly:2 word:14 onto:1 cannot:5 operator:1 applying:2 restriction:4 equivalent:1 map:1 missing:1 maximizing:10 annealed:1 layout:1 center:2 rectangular:1 assigns:1 rule:1 embedding:13 handle:1 hurt:2 analogous:1 updated:1 imagine:2 play:1 decode:1 origin:1 agreement:1 element:3 rumelhart:2 bottom:2 role:1 probabili:1 solved:1 ensures:1 connected:2 decrease:1 trained:1 girl:11 easily:2 represented:1 train:2 distinct:1 fast:2 effective:1 london:1 artificial:1 lift:1 encoded:1 stanford:1 solve:1 reconstruct:1 compressed:1 encoder:1 syntactic:1 think:1 noisy:3 jointly:1 final:1 differentiate:1 sequence:1 ucl:2 reconstruction:5 interaction:1 kzl:1 realization:1 margaret:1 getting:1 parent:1 zp:2 r1:12 produce:1 converges:1 help:2 depending:1 andrew:1 ac:1 fixing:1 completion:13 eq:4 solves:2 come:1 indicate:2 closely:1 correct:6 luckily:1 centered:1 vx:1 everything:1 implementing:1 assign:1 generalization:14 proposition:1 probable:1 secondly:1 hold:1 around:1 considered:1 mapping:2 pointing:1 iw:1 saw:1 minimization:1 mit:1 j7:1 always:6 gaussian:1 aim:1 encode:1 maria:1 indicates:2 sense:1 inference:1 rigid:1 suffix:1 relation:42 italian:4 iu:3 issue:1 among:4 noun:23 summed:1 ness:1 equal:4 once:2 never:1 having:4 identical:1 represents:1 marcello:1 few:4 composed:1 simultaneously:1 divergence:1 consisting:2 certainly:1 analyzed:1 mixture:1 nephew:1 held:1 subtrees:1 necessary:1 arthur:1 tree:47 euclidean:1 old:3 pollack:4 instance:5 column:1 goodness:3 queen:1 maximization:1 phrase:4 introducing:1 subset:2 hundred:1 uniform:5 father:3 successful:1 recognizing:2 erlbaum:1 too:1 answer:4 chooses:1 person:1 density:1 stay:1 probabilistic:3 decoding:4 squared:1 thesis:2 satisfied:1 containing:2 woman:2 charlotte:2 collapsing:1 admit:3 cognitive:1 caused:1 try:2 picked:1 root:6 reconstructor:3 analyze:1 doctor:2 parallel:1 contribution:7 minimize:1 square:1 formed:3 variance:3 kaufmann:1 maximized:1 correspond:2 generalize:5 decodes:1 multiplying:2 worth:1 published:2 explain:1 husband:1 definition:1 james:3 obvious:1 icml2000:1 associated:1 couple:1 stop:1 proved:1 logical:2 knowledge:3 back:1 appears:2 originally:1 higher:5 formulation:2 done:3 though:1 just:1 implicit:1 hand:2 christopher:2 indicated:1 believe:1 name:1 concept:38 requiring:1 assigned:3 symmetric:3 leibler:1 semantic:3 numerator:1 during:5 aunt:2 width:1 paccanaro:3 trying:1 evident:1 complete:11 christine:1 reasoning:1 charles:1 began:1 common:1 lre:22 volume:1 belong:1 association:1 mellon:1 imposing:1 mother:1 nationality:1 had:5 impressive:1 own:1 belongs:1 certain:3 binary:13 xe:1 raam:6 seen:1 morgan:1 greater:1 care:1 recognized:1 maximize:1 colin:1 relates:2 branch:1 violate:1 multiple:1 infer:5 match:2 alberto:7 divided:1 bigger:2 halt:1 involving:1 denominator:2 iteration:1 represent:5 normalization:1 c1:11 want:2 diagram:5 completes:1 else:1 extra:3 ascent:1 tend:1 member:2 grammatically:1 jordan:1 call:1 easy:1 psychology:1 architecture:3 competing:1 idea:1 whether:2 cause:2 enumerate:1 generally:1 clear:1 involve:1 amount:1 neocortex:1 gih:1 bac:1 category:1 mcclelland:2 generate:3 specifies:1 supplied:1 exist:2 notice:3 brother:4 sign:1 neuroscience:1 delta:1 correctly:5 sister:3 per:2 disjoint:1 write:2 discrete:2 shall:8 carnegie:1 group:4 four:1 threshold:1 neither:1 kept:1 graph:1 pietro:3 merely:1 sum:1 wife:1 i5:1 striking:1 arrive:1 family:19 almost:8 decide:1 layer:2 completing:1 followed:1 encountered:1 annual:1 adapted:1 constraint:3 encodes:1 generates:1 performing:1 separable:1 format:1 department:2 numbering:2 according:4 leabra:1 poor:2 conjugate:1 belonging:2 son:2 making:2 randall:1 outlier:2 gradually:1 pr:1 previously:2 jennifer:1 discus:1 know:4 end:1 operation:2 gaussians:5 victoria:1 hierarchical:3 away:1 appropriate:3 enforce:1 appearing:1 original:1 top:3 remaining:1 include:1 denotes:1 completed:2 quinlan:2 exploit:1 build:2 society:1 noticed:1 added:1 gradient:2 distance:1 valence:2 separate:1 link:1 entity:2 concatenation:1 w0:2 argue:1 trivial:2 assuming:1 code:4 modeled:1 relationship:3 providing:1 minimizing:1 unrolled:1 unfortunately:1 negative:3 daughter:2 proper:1 allowing:1 observation:1 coincident:1 pat:1 situation:1 hinton:10 relational:8 extended:1 pdp:1 arbitrary:1 verb:13 david:2 pair:5 cast:1 dog:3 sentence:17 learned:6 able:16 bar:3 beyond:1 usually:1 pattern:6 xm:1 eighth:1 adjective:12 built:3 critical:1 suitable:1 representing:4 altered:1 auto:2 extract:1 prior:1 checking:1 multiplication:2 relative:1 fully:1 generation:1 limitation:1 geoffrey:4 validation:6 sufficient:2 editor:2 cd:1 foil:3 row:8 elsewhere:1 placed:4 last:1 english:3 ugly:2 distributed:24 grammatical:2 tolerance:1 dimension:2 vocabulary:2 giovanni:2 overcome:1 author:2 made:2 far:1 facsimile:1 reconstructed:4 compact:2 kullback:1 tuples:1 discriminative:3 don:4 triplet:46 pretty:4 why:1 learn:9 uncle:3 excellent:1 domain:2 constituted:3 linearly:1 aurelio:1 terminated:1 noise:1 nothing:1 allowed:1 child:2 fig:9 fashion:1 gatsby:2 fails:1 inferring:1 decoded:5 sub:4 position:9 concatenating:1 third:4 learns:3 young:2 specific:1 showing:1 arity:3 symbol:5 list:2 r2:4 burden:1 exists:2 adding:2 effectively:2 phd:2 commutative:1 simply:1 expressed:1 ordered:2 contained:3 compressor:5 applies:1 corresponds:1 determines:1 satisfies:1 goal:2 sized:2 called:5 total:2 isomorphic:1 rarely:1 internal:2 people:8 incorporate:1 tested:1
1,172
2,069
Thin Junction Trees Francis R. Bach Computer Science Division University of California Berkeley, CA 94720 fbach@cs.berkeley.edu Michael I. Jordan Computer Science and Statistics University of California Berkeley, CA 94720 jordan@cs.berkeley.edu Abstract We present an algorithm that induces a class of models with thin junction trees?models that are characterized by an upper bound on the size of the maximal cliques of their triangulated graph. By ensuring that the junction tree is thin, inference in our models remains tractable throughout the learning process. This allows both an efficient implementation of an iterative scaling parameter estimation algorithm and also ensures that inference can be performed efficiently with the final model. We illustrate the approach with applications in handwritten digit recognition and DNA splice site detection. Introduction Many learning problems in complex domains such as bioinformatics, vision, and information retrieval involve large collections of interdependent variables, none of which has a privileged status as a response variable or class label. In such problems, the goal is generally that of characterizing the principal dependencies in the data, a problem which is often cast within the framework of multivariate density estimation. Simple models are often preferred in this setting, both for their computational tractability and their relative immunity to overfitting. Thus models involving low-order marginal or conditional probabilities? e.g., naive independence models, trees, or Markov models?are in wide use. In problems involving higher-order dependencies, however, such strong assumptions can be a serious liability. A number of methods have been developed for selecting models of higher-order dependencies in data, either within the maximum entropy setting?in which features are selected [9, 16]?and the graphical model setting?in which edges are selected [8]. Simplicity also plays an important role in the design of these algorithms; in particular, greedy methods that add or subtract a single feature or edge at a time are generally employed. The model that results at each step of this process, however, is often not simple, and this is problematic both computationally and statistically in large-scale problems. In the current paper we describe a methodology that can be viewed as a generalization of the Chow-Liu algorithm for constructing tree models [2]. Note that tree models have the property that their junction trees have no more than two nodes in any clique?the treewidth of tree models is one. In our generalization, we allow the treewidth to be a larger, but still controlled, value. We fit data within the space of models having ?thin? junction trees. Models with thin junction trees are tractable for exact inference, indeed the complexity of any type of inference (joint, marginal, conditional) is controlled by the upper bound that is imposed on the treewidth. This makes it possible to achieve some of the flexibility that is often viewed as a generic virtue of generative models, but is not always achievable in practice. For example, in the classification setting we are able to classify partially observed data (e.g., occluded digits) in a simple and direct way?we simply marginalize away the unobserved variables, an operation which is tractable in our models. We illustrate this capability in a study of handwritten digit recognition in Section 4.2, where we compare thin junction trees and support vector machines (SVMs), a discriminative technique which does not come equipped with a simple and principled method for handling partially observed data. As we will see, thin junction trees are quite robust to missing data in this domain. There are a number of issues that need to be addressed in our framework. In particular, tree models come equipped with particularly efficient algorithms for parameter estimation and model selection?algorithms which do not generalize readily to non-tree models, including thin junction tree models. It is important to show that efficient algorithms can nonetheless be found to fit such models. We show how this can be achieved in Sections 1, 2 and 3. Empirical results using these algorithms are presented in Section 4. 1 Feature induction   We assume an input space with variables and a target probability distribution . Our goal is to find a probability distribution that minimizes the Kullback-Leibler divergence . Consider a vector-valued ?feature? or ?sufficient statistic? , where is the dimensionality of the feature space. The feature can also be thought in terms of its components as a set of real-valued features . We focus on exponential family distributions (also known as ?Gibbs? or ?maximum entropy? distributions) based on these features: where is a parameter vector, is a base-measure (typically uniform), and is the normalizing constant. (Section 3 considers the closely-related problem of inducing edges rather than features).       #         !#"!$ & %'   ( ()+* 8:9:;=<>;??+?@; BA *   %  ,%  .- /    02143657 Each feature is a function of a certain subset of variables, and we let index the subset of variables referred to by feature . Let us consider the undirected graphical model , where the set of edges is the set of all pairs are the maximal cliques of the graph included in at least one . With this definition the and, if is decomposable in this graph, the exponential family distribution with features and reference distribution is also decomposable in this graph. We assume without loss of generality that the graph is connected. For each possible triangulation of the graph, we can define a junction tree [4], where for all there exists a maximal clique containing . The complexity of exact inference depends on the size of the maximal clique of the triangulated graph. We define the treewidth of our original graph to be one less than the minimum possible value of this maximal clique size for all possible triangulations. We say that a graphical model has a thin junction tree if its treewidth is small. 01 # 0G1 CD  5 (; E + 01 E :1 H F H Our basic feature induction algorithm is a constrained variant of that proposed by [9]. Given a set of available features, we perform a greedy search to find the set of features that enables the best possible fit to , under the constraint of having a thin junction tree. At each step, candidates are ranked according to the gain in KL divergence, with respect to the empirical distribution, that would be achieved by their addition to the current set of features. Features that would generate a graphical cover with treewidth greater than a given upper bound are removed from the ranking.  % H The parameter values are held fixed during each step of the feature ranking process. Once a set of candidate features are chosen, however, we reestimate all of the parameters (using the algorithm to be described in Section 2) and iterate. F EATURE I NDUCTION    . ;%    , , a set of available features 1. Initialization: 2. Repeat steps (a) to (d) until no further progress is made with respect to a model selection criterion (e.g., MDL or cross-validation) (a) Ranking: generate samples from and rank feature candidates according to the KL gain (b) Elimination: remove all candidates that would generate a model with treewidth greater than  and add them to (c) Selection: select the  best features  (d) Parameter Estimation: Estimate using the junction tree implementation of Iterative Scaling (see Section 2)  H % ;?+??#; Freezing the parameters during the feature ranking step is suboptimal, but it yields an essential computational efficiency. In particular, as shown by [9], under these conditions we can rank a new feature by solving a polynomial equation whose degree is the number of values can take minus one, and whose coefficients are expectations under of functions of . This equation has only one root and can be solved efficiently by Newton?s method. When the feature is binary the process is even more efficient?the equation is linear and can be solved directly. Consequently, with a single set of samples from , we can rank many features very cheaply.   For the feature elimination operation, algorithms exist that determine in time linear in the number of nodes whether a graph has a treewidth smaller than , and if so output a triangulation in which all cliques are of size less than [1]. These algorithms are super-exponential in , however, and thus are applicable only to problems with small treewidths. In practice we have had success using fast heuristic triangulation methods [11] that allow us to guarantee the existence of a junction tree with a maximal clique no larger than for a given model. (This is a conservative technique that may occasionally throw out models that in fact have small treewidth). H H H H A critical bottleneck in the algorithm is the parameter estimation step, and it is important to develop a parameter estimation algorithm that exploits the bounded treewidth property. We now turn to this problem. 2 Iterative Scaling using the junction tree Fitting an exponential family distribution under expectation constraints is a well studied problem; the basic technique is known as Iterative Scaling. A generalization of Iterative Proportional Fitting (IPF), it updates the parameters sequentially [5]. Algorithms that update the parameters in parallel have also been proposed; in particular the Generalized Iterative Scaling algorithm [6], which imposes the constraint that the features sum to one, and the Improved Iterative Scaling algorithm [9], which removes this constraint. These algorithms have an important advantage in our setting in that, for each set of parameter updates, they only require computations of expectation that can all be estimated with a single set of samples from the current distribution. % When the input dimensionality is large, however, we would like to avoid sampling algorithms altogether. To do so we exploit the bounded treewidth of our models. We present a novel algorithm that uses the junction tree and the structure of the problem to speed up parameter estimation. The algorithm generalizes to Gibbs distributions the ?effective IPF? algorithm of [10]. When working with a junction tree, a efficient way of performing Iterative Scaling is to update parameters block by block so that each update is performed for a relatively small number of features on a small number of variables. Each block can be fit with any parameter estimation algorithm, in particular Improved Iterative Scaling (IIS). The following algorithm exploits this idea by grouping the features whose supports are in the same clique of the triangulated graph. Thus, parameter estimation is done in spaces of dimensions at most , and all the needed expectations can be evaluated cheaply. H 9 2.1 Notation   &        1 ;+??+?#; : 11  Let be our -dimensional feature. Let denote the maximal cliques of the triangulated graph, with potentials . We assign each feature to one of the cliques that contains . For each clique we denote as the set of features assigned to .    G0  1 2.2 Algorithm E FFICIENT I TERATIVE S CALING 80 1   1 A 1. Initialization: supp ?Construct a junction tree associated with the subsets ?Assign each to one , such that (equivalent to determining for all )  ?Set and decompose onto the junction tree ?Set  1  ;?+??#; 1  1 %  ,% ;?+??; %          01 3        % 2. Loop until convergence: Repeat step (3) until convergence of the ?s  3. Loop through all cliques: Repeat steps (a) to (c) for all cliques  (a) Define the root of the junction tree to be (b) Collect evidence from the leaves to the root of the junction tree and normalize potential (c) Calculate the maximum likelihood -dimensional exponential family distribution with features and reference distribution , using IIS. Replace by this distribution and add the resulting parameters (one for each feature in ) to the corresponding ?s: .            %  &% 1  ; ?+??+; % 1     After step (b), the potential is exactly marginalized to , so that performing IIS for can be done using instead of the full distribution . Moreover, each the features pass through all the cliques is equivalent to one pass of Iterative Scaling and therefore this algorithm converges to the maximum likelihood distribution.   3 Edge induction Thus far we have emphasized the exponential family representation. Our algorithm can, however, be adapted readily to the problem of learning the structure of a graphical model. This is achieved by using features that are indicators of subsets of variables, ensuring that there is one such indicator for every combination of values of the variables in a clique. In this case, Iterative Scaling reduces to Iterative Proportional Fitting. We generally employ a further approximation when ranking and selecting edges. In particular, we evaluate an edge only in terms of the two variables associated directly with the edge. The clique formed by the addition of the edge, however, may involve additional higher-order dependencies, which can be parameterized and incorporated in the model. Evaluating edges in this way thus underestimates the potential gain in KL divergence. 20 15 10 5 0 0 10 20 30 Figure 1: (Left) Circular Boltzmann machine of treewidth 4. (Right) Proportion (in ) of edges not in common between the fitted model and the generating model vs the number of available training examples (in thousands). We should not expect to be able to find an exact edge-selection method?recent work by Srebro [15] has shown that the related problem of finding the maximum likelihood graphical model with bounded treewidth is NP-hard. 4 Empirical results 4.1 Small graphs with known generative model In this experiment we generate samples from a known graphical model and fit our model to the data. We consider circular Boltzmann machines of known treewidth equal to 4 as shown in Figure 1. Our networks all have 32 nodes and the weights were selected from a        ?so that each edge is significant. For an increasing uniform distribution in  number of training samples, ten replications were performed for each case using our feature induction algorithm with maximum treewidth equal to 4. Figure 1 shows that with enough samples we are able to recover the structure almost exactly (up to of the original edges). < 9 9 =< ? 4.2 MNIST digit dataset In this section we study the performance of the thin junction tree method on the MNIST dataset of handwritten digits. While discriminative methods outperform generative methods in this high-dimensional setting [12], generative methods offer capabilities that are not provided by discriminative classifiers; in particular, the ability to deal with large fractions of missing pixels and the ability to to reconstruct images from partial data. It is of interest to see how much performance loss we incur and how much robustness we gain by using a sophisticated generative model for this problem. 9 9 <  <  The MNIST training set is composed of 4-bit grayscale pixels that have been   resized and cropped to binary images (an example is provided in the leftmost plots in Figure 2). We used thin junction trees as density estimators in the 256-dimensional pixel space by training ten different models, one for each of the ten classes. We used binary  features of the form . No vision-based techniques such as de-skewing or virtual examples were used. We utilized ten percent fractions of the training data for crossvalidation and test.      9 Density estimation: The leftmost plot in Figure 3 shows how increasing the maximal allowed treewidth, ranging from 1 (trees) to 15, enables a better fit to data. Classification: We built classifiers from the bank of ten thin junction tree (?TJT?) models using one of the following strategies: (1) take the maximum likelihood among the ten Figure 2: Digit from the MNIST database. From left to right, original digit, cropped and resized digits used in our experiments, 50% of missing values, 75% of missing values, occluded digit. 100 70 80 65 60 60 40 55 50 20 0 5 10 15 0 0 50 100 Figure 3: (Left) Negative log likelihood for the digit 2 vs maximal allowed treewidth. (Right) Error rate as a function of the percentage of erased pixels for the TJT classifier (plain) and a support vector machine (dotted). See text for details. models (TJT-ML), or (2) train a discriminative model using the outputs of the ten models. We used softmax regression (TJT-Softmax) and the support vector machine (TJT-SVM) in the latter case. The classification error rates were as follows: LeNet 0.7, SVM 0.8, Product of experts, 2.0, TJT-SVM 3.8, TJT-Softmax 4.2, TJT-ML 5.3, Chow-Liu 8.5, and Linear classifier 12.0. (See [12] and [13] for further details on the non-TJT models). It is important to emphasize that our models are tractable for full joint inference; indeed, the junction trees have a maximal clique size of 10 in the largest models we used on the ten classes. Thus we can use efficient exact calculations to perform inference. The following two sections demonstrate the utility of this fact. Missing pixels: We ran an experiment in which pixels were chosen uniformly at random and erased, as shown in Figure 2. In our generative model, we treat them as hidden variables that were marginalized out. The rightmost plot in Figure 3 shows the error rate on the testing set as a function of the percentage of unknown pixels, for our models and for a SVM. In the case of the SVM, we used a polynomial kernel of degree four [7] and we tried various heuristics to fill in the value of the non-observed pixels, such as the average of that pixel over the training set or the value of a blank pixel. Best classification performance was achieved with replacing the missing value by the value of a blank pixel. Note that very little performance decrement is seen for our classifier even with up to 50 percent of the pixels missing, while for the SVM, although performance is better for small percentages, performance degrades more rapidly as the percentage of erased digits increases. Reconstruction: We conducted an additional experiment in which the upper halves of images were erased. We ran the junction tree inference algorithm to fill in these missing values, choosing the maximizing value of the conditional probability (max-propagation). Figure 3 shows the results. For each line, from left to right, we show the original digit, the digit after erasure, reconstructions based on the model having the maximum likelihood, and 0 0 2 6 5 5 8 3 1 1 9 7 6 6 0 2 2 2 6 5 7 7 9 4 3 3 5 8 8 8 6 3 4 4 7 9 9 7 9 3 Figure 4: Reconstructions of images whose upper halves have been deleted. See text for details. reconstruction based on the model having the second and third largest values of likelihood. 4.3 SPLICE Dataset The task in this dataset is to classify splice junctions in DNA sequences. Splice junctions can either be an exon/intron (EI) boundary, an intron/exon (IE) boundary, or no boundary. (Introns are the portions of genes that are spliced out during transcription; exons are retained in the mRNA). Each sample is a sequence of 60 DNA bases (where each base can take one of four values, A,G,C, or T). The three different classes are: EI exactly at the middle (between the 30th and the 31st bases), IE exactly at the middle (between the 30th and the 31st bases), no splice junction. The dataset is composed of 3175 training samples. In order to be able to compare to previous experiments using this dataset, performance is assessed by picking 2000 training data points at random and testing on the 1175 others, with 20 replications. We treat classification as a density estimation problem in this case by treating the class variable as another variable. We classify by choosing the value of that maximizes the conditional probability . We tested both feature induction and edge induction; in the  were former case only binary features that are products of features of the form tested and induced. MDL was used to pick the number of features or edges.      ?  ?      ?9 Our feature induction algorithm, with a maximum treewidth equal to 5, gave an error rate of , while the edge induction algorithm gave an error rate of . This is better than the best reported results in the literature; in particular, neural networks have an error rate of and the Chow and Liu algorithm has an error rate of [14].  ? 5 Conclusions We have described a methodology for feature selection, edge selection and parameter estimation that can be viewed as a generalization of the Chow-Liu algorithm. Drawing on the feature selection methods of [9, 16], our method is quite general, building an exponential family model from the general vocabulary of features on overlapping subsets of variables. By maintaining tractability throughout the learning process, however, we build this flexible representation of a multivariate density while retaining many of the desirable aspects of the Chow-Liu algorithm. Our methodology applies equally well to feature or edge selection. In large-scale, sparse domains in which overfitting is of particular concern, however, feature selection may be the preferred approach, in that it provides a finer-grained search in the space of simple models than is allowed by the edge selection approach. Acknowledgements We wish to acknowledge NSF grant IIS-9988642 and ONR MURI N00014-00-1-0637. The results presented here were obtained using Kevin Murphy?s Bayes Net Matlab toolbox and SVMTorch [3]. References [1] H. Bodlaender, A linear-time algorithm for finding tree-decompositions of small treewidth, Siam J. Computing, 25, 105-1317, 1996. [2] C.K. Chow and C.N. Liu, Approximating discrete probability distributions with dependence trees, IEEE Trans. Information Theory, 42, 393-405, 1990. [3] R. Collobert and S. Bengio, SVMTorch: support vector machines for large-scale regression problems, Journal of Machine Learning Research, 1, 143-160, 2001. [4] R.G. Cowell, A.P. Dawid, S.L. Lauritzen, and D.J. Spiegelhalter, Probabilistic Networks and Expert Systems, Springer-Verlag, New York, 1999. [5] I. Csisz?ar, I-divergence geometry of probability distributions and minimization problems, Annals of Probability, 3, 146-158, 1975. [6] J.N. Darroch and D. Ratcliff, Generalized iterative scaling for log-linear models, Ann. Math. Statist., 43, 1470-1480, 1972. [7] D. DeCoste and B. Sch?olkopf, Training invariant support vector machines, Machine Learning, 46, 1-3, 2002. [8] D. Heckerman, D. Geiger, and D.M. Chickering, Learning Bayesian networks: The combination of knowledge and statistical data, Machine Learning, 20, 197-243, 1995. [9] S. Della Pietra, V. Della Pietra, and J. Lafferty, Inducing features of random fields, IEEE Trans. PAMI, 19, 380-393, 1997. [10] R. Jirousek and S. Preucil, On the effective implementation of the iterative proportional fitting procedure, Computational Statistics and Data Analysis, 19, 177-189, 1995. [11] U. Kjaerulff, Triangulation of graphs?algorithms giving small total state space, Technical Report R90-09, Dept. of Math. and Comp. Sci., Aalborg Univ., Denmark, 1990. [12] Y. Le Cun, http://www.research.att.com/?yann/exdb/mnist/index.html [13] G. Mayraz and G. Hinton, Recognizing hand-written digits using hierarchical products of experts, Adv. NIPS 13, MIT Press, Cambridge, MA, 2001. [14] M. Meila and M.I. Jordan, Learning with mixtures of trees, Journal of Machine Learning Research, 1, 1-48, 2000. [15] N. Srebro, Maximum likelihood bounded tree-width Markov networks, in UAI 2001. [16] S.C. Zhu, Y.W. Wu, and D. Mumford, Minimax entropy principle and its application to texture modeling, Neural Computation, 9, 1997.
2069 |@word middle:2 polynomial:2 achievable:1 proportion:1 tried:1 decomposition:1 pick:1 minus:1 liu:6 contains:1 att:1 selecting:2 rightmost:1 current:3 blank:2 com:1 mayraz:1 written:1 readily:2 enables:2 remove:2 plot:3 treating:1 update:5 v:2 greedy:2 selected:3 generative:6 leaf:1 half:2 provides:1 math:2 node:3 direct:1 replication:2 fitting:4 indeed:2 little:1 decoste:1 equipped:2 increasing:2 provided:2 bounded:4 notation:1 moreover:1 maximizes:1 minimizes:1 developed:1 unobserved:1 finding:2 guarantee:1 berkeley:4 every:1 exactly:4 classifier:5 grant:1 treat:2 pami:1 initialization:2 studied:1 collect:1 statistically:1 testing:2 practice:2 block:3 digit:14 procedure:1 erasure:1 empirical:3 thought:1 onto:1 marginalize:1 selection:10 www:1 equivalent:2 imposed:1 missing:8 maximizing:1 mrna:1 simplicity:1 decomposable:2 estimator:1 fill:2 annals:1 target:1 play:1 exact:4 us:1 dawid:1 recognition:2 particularly:1 utilized:1 muri:1 database:1 jirousek:1 observed:3 role:1 solved:2 calculate:1 thousand:1 ensures:1 connected:1 adv:1 spliced:1 removed:1 ran:2 principled:1 complexity:2 occluded:2 solving:1 incur:1 division:1 efficiency:1 exon:3 joint:2 various:1 train:1 univ:1 fast:1 describe:1 effective:2 kevin:1 choosing:2 quite:2 whose:4 larger:2 valued:2 heuristic:2 say:1 drawing:1 reconstruct:1 ability:2 statistic:3 g1:1 final:1 advantage:1 sequence:2 net:1 reconstruction:4 maximal:10 product:3 loop:2 rapidly:1 flexibility:1 achieve:1 inducing:2 csisz:1 normalize:1 olkopf:1 crossvalidation:1 convergence:2 generating:1 converges:1 illustrate:2 develop:1 lauritzen:1 progress:1 strong:1 throw:1 c:2 treewidth:19 come:2 triangulated:4 closely:1 elimination:2 virtual:1 require:1 assign:2 generalization:4 decompose:1 tjt:9 estimation:12 applicable:1 label:1 largest:2 minimization:1 mit:1 always:1 super:1 rather:1 avoid:1 resized:2 focus:1 rank:3 likelihood:8 ratcliff:1 inference:8 typically:1 chow:6 hidden:1 pixel:12 issue:1 classification:5 among:1 flexible:1 html:1 retaining:1 constrained:1 softmax:3 marginal:2 equal:3 once:1 construct:1 having:4 field:1 sampling:1 thin:13 np:1 others:1 report:1 serious:1 employ:1 aalborg:1 composed:2 divergence:4 murphy:1 pietra:2 geometry:1 detection:1 interest:1 circular:2 mdl:2 mixture:1 held:1 edge:20 partial:1 tree:37 fitted:1 classify:3 modeling:1 cover:1 ar:1 tractability:2 subset:5 uniform:2 recognizing:1 conducted:1 reported:1 dependency:4 st:2 density:5 siam:1 ie:2 probabilistic:1 picking:1 michael:1 containing:1 expert:3 supp:1 potential:4 de:1 coefficient:1 ranking:5 depends:1 collobert:1 performed:3 root:3 francis:1 portion:1 recover:1 bayes:1 capability:2 parallel:1 formed:1 efficiently:2 yield:1 generalize:1 handwritten:3 bayesian:1 none:1 comp:1 finer:1 reestimate:1 definition:1 underestimate:1 nonetheless:1 associated:2 gain:4 dataset:6 knowledge:1 dimensionality:2 sophisticated:1 higher:3 methodology:3 response:1 improved:2 done:2 evaluated:1 generality:1 until:3 working:1 hand:1 freezing:1 replacing:1 ei:2 overlapping:1 propagation:1 ipf:2 building:1 former:1 lenet:1 assigned:1 leibler:1 deal:1 during:3 width:1 criterion:1 generalized:2 leftmost:2 exdb:1 demonstrate:1 percent:2 image:4 ranging:1 novel:1 common:1 significant:1 cambridge:1 gibbs:2 meila:1 had:1 add:3 base:5 multivariate:2 r90:1 recent:1 triangulation:5 occasionally:1 certain:1 n00014:1 verlag:1 binary:4 success:1 onr:1 seen:1 minimum:1 greater:2 additional:2 employed:1 determine:1 ii:4 full:2 desirable:1 reduces:1 technical:1 characterized:1 calculation:1 bach:1 cross:1 retrieval:1 offer:1 equally:1 privileged:1 controlled:2 ensuring:2 involving:2 basic:2 variant:1 regression:2 vision:2 expectation:4 kernel:1 achieved:4 addition:2 cropped:2 addressed:1 sch:1 fbach:1 induced:1 undirected:1 lafferty:1 jordan:3 bengio:1 enough:1 iterate:1 independence:1 fit:6 gave:2 suboptimal:1 idea:1 bottleneck:1 whether:1 utility:1 darroch:1 york:1 skewing:1 matlab:1 generally:3 involve:2 ten:8 statist:1 induces:1 svms:1 dna:3 generate:4 http:1 outperform:1 exist:1 percentage:4 problematic:1 nsf:1 dotted:1 estimated:1 discrete:1 four:2 deleted:1 graph:13 fraction:2 sum:1 parameterized:1 throughout:2 family:6 almost:1 yann:1 wu:1 geiger:1 scaling:11 bit:1 bound:3 adapted:1 constraint:4 aspect:1 speed:1 performing:2 relatively:1 according:2 combination:2 smaller:1 heckerman:1 cun:1 invariant:1 computationally:1 equation:3 remains:1 turn:1 needed:1 tractable:4 junction:29 operation:2 available:3 generalizes:1 hierarchical:1 away:1 generic:1 robustness:1 altogether:1 bodlaender:1 existence:1 original:4 graphical:7 marginalized:2 newton:1 maintaining:1 exploit:3 giving:1 build:1 approximating:1 g0:1 mumford:1 strategy:1 degrades:1 dependence:1 sci:1 evaluate:1 considers:1 induction:8 denmark:1 index:2 retained:1 negative:1 ba:1 implementation:3 design:1 boltzmann:2 unknown:1 perform:2 upper:5 markov:2 acknowledge:1 hinton:1 incorporated:1 treewidths:1 cast:1 pair:1 kl:3 toolbox:1 immunity:1 eature:1 california:2 nip:1 trans:2 able:4 built:1 including:1 max:1 critical:1 ranked:1 indicator:2 zhu:1 minimax:1 spiegelhalter:1 naive:1 text:2 interdependent:1 literature:1 acknowledgement:1 determining:1 relative:1 loss:2 expect:1 proportional:3 srebro:2 validation:1 kjaerulff:1 degree:2 sufficient:1 imposes:1 principle:1 bank:1 cd:1 repeat:3 liability:1 allow:2 wide:1 characterizing:1 sparse:1 boundary:3 dimension:1 plain:1 evaluating:1 vocabulary:1 collection:1 made:1 far:1 emphasize:1 preferred:2 status:1 kullback:1 gene:1 clique:18 ml:2 transcription:1 overfitting:2 sequentially:1 uai:1 discriminative:4 grayscale:1 search:2 iterative:14 terative:1 robust:1 ca:2 complex:1 constructing:1 domain:3 decrement:1 allowed:3 site:1 referred:1 wish:1 exponential:7 candidate:4 chickering:1 third:1 grained:1 splice:5 emphasized:1 intron:3 svm:6 virtue:1 normalizing:1 grouping:1 exists:1 essential:1 evidence:1 mnist:5 concern:1 texture:1 subtract:1 entropy:3 simply:1 cheaply:2 partially:2 applies:1 cowell:1 springer:1 ma:1 conditional:4 goal:2 viewed:3 consequently:1 ann:1 replace:1 erased:4 hard:1 included:1 uniformly:1 principal:1 conservative:1 total:1 pas:2 select:1 support:6 latter:1 assessed:1 bioinformatics:1 svmtorch:2 dept:1 tested:2 della:2 handling:1
1,173
207
524 Fablman and Lebiere The Cascade-Correlation Learning Architecture Scott E. Fahlman and Christian Lebiere School of Computer Science Carnegie-Mellon University Pittsburgh, PA 15213 ABSTRACT Cascade-Correlation is a new architecture and supervised learning algorithm for artificial neural networks. Instead of just adjusting the weights in a network of fixed topology. Cascade-Correlation begins with a minimal network, then automatically trains and adds new hidden units one by one, creating a multi-layer structure. Once a new hidden unit has been added to the network, its input-side weights are frozen. This unit then becomes a permanent feature-detector in the network, available for producing outputs or for creating other, more complex feature detectors. The Cascade-Correlation architecture has several advantages over existing algorithms: it learns very quickly, the network .determines its own size and topology, it retains the structures it has built even if the training set changes, and it requires no back-propagation of error signals through the connections of the network. 1 DESCRIPTION OF CASCADE?CORRELATION The most important problem preventing the widespread application of artificial neural networks to real-world problems is the slowness of existing learning algorithms such as back-propagation (or "backprop"). One factor contributing to that slowness is what we call the moving target problem: because all of the weights in the network are changing at once, each hidden units sees a constantly changing environment. Instead of moving quickly to assume useful roles in the overall problem solution, the hidden units engage in a complex dance with much wasted motion. The Cascade-Correlation learning algorithm was developed in an attempt to solve that problem. In the problems we have examined, it learns much faster than back-propagation and solves some other problems as well. The Cascade-Correlation Learning Architecture Outputs o 0 Output Units Hidden Unit 2 Hidden unit 1 ~~--------~~--.---- o--------~&_-------mr_--------------~----~--~~ Inpu~ O--------~~----~H-------------~.---~~--~ o--------~~------~--------------------------~ +1 Figure 1: The Cascade architecture, after two hidden units have been added. The vertical lines sum all incoming activation. Boxed connections are frozen, X connections are trained repeatedly. Cascade-Correlation combines two key ideas: The first is the cascade architecture, in which hidden units are added to the network one at a time and do not change after they have been added. The second is the learning algorithm, which creates and installs the new hidden units. For each new hidden unit, we attempt to maximize the magnitude of the correlation between the new unit's output and the residual error signal we are trying to eliminate. The cascade architecture is illustrated in Figure 1. It begins with some inputs and one or more output units, but with no hidden units. The number of inputs and outputs is dictated by the problem and by the I/O representation the experimenter has chosen. Every input is connected to every output unit by a connection with an adjustable weight. There is also a bias input, permanently set to +1. The output units may just produce a linear sum of their weighted inputs, or they may employ some non-linear activation function. In the experiments we have run so far, we use a symmetric sigmoidal activation function (hyperbolic tangent) whose output range is -1.0 to + 1.0. For problems in which a precise analog output is desired, instead of a binary classification, linear output units might be the best choice, but we have not yet studied any problems of this kind. We add hidden units to the network one by one. Each new hidden unit receives a connection from each of the network's original inputs and also from every pre-existing hidden unit. The hidden unit's input weights are frozen at the time the unit is added to the net; only the output connections are trained repeatedly. Each new unit therefore adds 525 526 Fahlman and Lebiere a new one-unit "layer" to the network, unless some of its incoming weights happen to be zero. This leads to the creation of very powerful high-order feature detectors; it also may lead to very deep networks and high fan-in to the hidden units. There are a number of possible strategies for minimizing the network depth and fan-in as new units are added, but we have not yet explored these strategies. The learning algorithm begins with no hidden units. The direct input-output connections are trained as well as possible over the entire training set. With no need to back-propagate through hidden units, we can use the Widrow-Hoff or "delta" rule, the Perceptron learning algorithm, or any of the other well-known learning algorithms for single-layer networks. In our simulations, we use Fahlman's "quickprop" algorithm [Fahlman, 1988] to train the output weights. With no hidden units, this acts essentially like the delta rule, except that it converges much faster. At some point, this training will approach an asymptote. When no significant error reduction has occurred after a certain number of training cycles (controlled by a "patience" parameter set by the operator), we run the network one last time over the entire training set to measure the error. If we are satisfied with the network's performance, we stop; if not, we attempt to reduce the residual errors further by adding a new hidden unit to the network. The unit-creation algorithm is described below. The new unit is added to the net, its input weights are frozen, and all the output weights are once again trained using quickprop. This cycle repeats until the error is acceptably small (or until we give up). To create a new hidden unit, we begin with a candidate unit that receives trainable input connections from all of the network's external inputs and from all pre-existing hidden units. The output of this candidate unit is not yet connected to the active network. We run a number of passes over the examples of the training set, adjusting the candidate unit's input weights after each pass. The goal of this adjustment is to maximize S, the sum over all output units 0 of the magnitude of the correlation (or, more precisely, the covariance) between V, the candidate unit's value, and Eo, the residual output error observed at unit o. We define S as S= L: L:(Vp o V) (Ep,o - Eo) p where 0 is the network output at which the error is measured and p is the training pattern. The quantities V and Eo are the values of V and Eo averaged over all patterns. In order to maximize S, we must compute 8Sj8wi, the partial derivative of S with respect to each of the candidate unit's incoming weights, Wi. In a manner very similar to the derivation of the back-propagation rule in [Rumelhart, 1986], we can expand and differentiate the fonnula for S to get 8Sj8Wj =L: uo(Ep,o - Eo)J;,lj,p p,o where U o is the sign of the correlation between the candidate's value and output o,ff, is The Cascade-Correlation Learning Architecture the derivative for pattern p of the candidate unit's activation function with respect to the sum of its inputs, and li,p is the input the candidate unit receives from unit i for pattern p. After computing 8 S/ 8Wi for each incoming connection, we can perform a gradient ascent to maximize S. Once again we are training only a single layer of weights. Once again we use the quickprop update rule for faster convergence. When S stops improving, we install the new candidate as a unit in the active network, freeze its input weights, and continue the cycle as described above. Because of the absolute value in the formula for S, a candidate unit cares only about the magnitude of its correlation with the error at a given output, and not about the sign of the correlation. As a rule, if a hidden unit correlates positively with the error at a given unit, it will develop a negative connection weight to that unit, attempting to cancel some of the error; if the correlation is negative, the output weight will be positive. Since a unit's weights to different outputs may be of mixed sign, a unit can sometimes serve two purposes by developing a positive correlation with the error at one output and a negative correlation with the error at another. Instead of a single candidate unit. it is possible to use a pool of candidate units, each with a different set of random initial weights. All receive the same input signals and see the same residual error for each pattern and each output. Because they do not interact with one another or affect the active network during training, all of these candidate units can be trained in parallel; whenever we decide that no further progress is being made, we install the candidate whose correlation score is the best. The use of this pool of candidates is beneficial in two ways: it greatly reduces the chance that a useless unit will be permanently installed because an individual candidate got stuck during training, and (on a parallel machine) it can speed up the training because many parts of weight-space can be explored simultaneously. The hidden and candidate units may all be of the same type, for example with a sigmoid activation function. Alternatively, we might create a pool of candidate units with a mixture of nonlinear activation functions-some sigmoid, some Gaussian, some with radial activation functions. and so on-and let them compete to be chosen for addition to the active network. To date, we have explored the all-sigmoid and all-Gaussian cases, but we do not yet have extensive simulation data on networks with mixed unit-types. One final note on the implementation of this algorithm: While the weights in the output layer are being trained, the other weights in the active network are frozen. While the candidate weights are being trained, none of the weights in the active network are changed. In a machine with plenty of memory. it is possible to record the unit-values and the output errors for an entire epoch, and then to use these cached values repeatedly during training. rather than recomputing them repeatedly for each training case. This can result in a tremendous speedup as the active network grows large. 527 528 Fahlman and Lebiere Figure 2: Training points for the two-spirals problem, and output pattern for one network trained with Cascade-Correlation. 2 BENCHMARK RESULTS 2.1 THE TWO-SPIRALS PROBLEM The "two-spirals" benchmark was chosen as the primary benchmark for this study because it is an extremely hard problem for algorithms of the back-propagation family to solve. n was first proposed by Alexis Wieland of MImE Corp. The net has two continuousvalued inputs and a single output. The training set consists of 194 X-Y values, half of which are to produce a +1 output and half a -1 output. These training points are arranged in two interlocking spirals that go around the origin three times, as shown in Figure 2a. The goal is to develop a feed-forward network with sigmoid units that properly classifies all 194 training cases. Some hidden units are obviously needed, since a single linear separator cannot divide two sets twisted together in this way. Wieland (unpublished) reported that a modified version of backprop in use at MITRE required 150,000 to 200,000 epochs to solve this problem, and that they had never obtained a solution using standard backprop. Lang and Witbrock [Lang, 1988] tried the problem using a 2-5-5-5-1 network (three hidden layers of five units each). Their network was unusual in that it provided "shortcut" connections: each unit received incoming connections from every unit in every earlier layer, not just from the immediately preceding layer. With this architecture, standard backprop was able to solve the problem in 20,000 epochs, backprop with a modified error function required 12,000 epochs, and quickprop required 8000. This was the best two-spirals performance reported to date. Lang and Witbrock also report obtaining a solution with a 2-5-5-1 net (only ten hidden units in all), but the solution required 60,000 quickprop epochs. We ran the problem 100 times with the Cascade-Correlation algorithm using a Sigmoidal activation function for both the output and hidden units and a pool of 8 candidate units. All trials were successful, requiring 1700 epochs on the average. (This number counts The Cascade-Correlation Learning Architecture both the epochs used to train output weights and the epochs used to train candidate units.) The number of hidden units built into the net varied from 12 to 19, with an average of 15.2 and a median of 15. Here is a histogram of the number of hidden units created: Hidden Units 12 13 14 15 16 17 18 19 Number of Trials 4 #### ######### 9 24 ######################## 19 ################### 24 ######################## 13 ############# 5 ##### 2 ## In terms of training epochs, Cascade-Correlation beats quickprop by a factor of 5 and standard backprop by a factor of 10, while building a network of about the same complexity (15 hidden units). In terms of actual computation on a serial machine, however, the speedup is much greater than these numbers suggest In backprop and quickprop, each training case requires a forward and a backward pass through all the connections in the network; Cascade-Correlation requires only a forward pass. In addition, many of the Cascade-Correlation epochs are run while the network is much smaller than its final size. Finally, the cacheing strategy described above makes it possible to avoid re-computing the unit values for parts of the network that are not changing. Suppose that instead of epochs, we measure learning time in connection crossings, defined as the number of multiply-accumulate steps necessary to propagate activation values forward through the network and error values backward. This measure leaves out some computational steps, but it is a more accurate measure of computational complexity than comparing epochs of different sizes or comparing runtimes on different machines. The Lang and Witbrock result of 20,000 backprop epochs requires about 1.1 billion connection crossings. Their solution using 8000 quickprop epochs on the same network requires about 438 million crossings. An average Cascade-Correlation run with a pool of 8 candidate units requires about 19 million crossings-a 23-fold speedup over quickprop and a 50-fold speedup over standard backprop. With a smaller pool of candidate units the speedup (on a serial machine) would be even greater, but the resulting networks might be somewhat larger. Figure 2b shows the output of a 12-hidden-unit network built by Cascade-Correlation as the input is scanned over the X-V field. This network properly classifies all 194 training points. We can see that it interpolates smoothly for about the first 1.5 turns of the spiral, but becomes a bit lumpy farther out, where the training points are farther apart. This "receptive field" diagram is similar to that obtained by Lang and Witbrock using backprop, but is somewhat smoother. 529 530 Fahlman and Lebiere 2.2 N-INPUT PARITY Since parity has been a popular benchmark among other researchers, we ran CascadeCorrelation on N-input parity problems with N ranging from 2 to 8. The best results were obtained with a sigmoid output unit and hidden units whose output is a Gaussian function of the sum of weighted inputs. Based on five trials for each value of N, our results were as follows: N Cases 2 3 4 5 6 7 8 4 8 16 32 64 128 256 Hidden Units 1 1 2 2-3 3 4-5 4-5 Average Epochs 24 32 66 142 161 292 357 For a rough comparison, Tesauro and Janssens [Tesauro, 1988] report that standard backprop takes about 2000 epochs for 8-input parity. In their study, they used 2N hidden units. Cascade-Correlation can solve the problem with fewer than N hidden units because it uses short-cut connections. As a test of generalization, we ran a few trials of Cascade-Correlation on the lO-input parity problem, training on either 50% or 25% of the 1024 patterns and testing on the rest. The number of hidden units built varied from 4 to 7 and training time varied from 276 epochs to 551. When trained on half of the patterns, perfonnance on the test set averaged 96% correct; when trained on one quarter of the patterns, test-set performance averaged 90% correct Note that the nearest neighbor algorithm would get almost all of the test-set cases wrong. 3 DISCUSSION We believe that that Cascade-Correlation algorithm offers the following advantages over network learning algorithms currently in use: ? There is no need to guess the size, depth, and connectivity pattern of the network in advance. A reasonably small (though not optimal) net is built automatically, perhaps with a mixture of unit-types . ? Cascade-Correlation learns fast In backprop, the hidden units engage in a complex dance before they settle into distinct useful roles; in Cascade-Correlation, each unit sees a fixed problem and can move decisively to solve that problem. For the problems we have investigated to date, the learning time in epochs grows roughly as NlogN, where N is the number of hidden units ultimately needed to solve the problem. The Cascade-Correlation Learning Architecture ? Cascade-Correlation can build deep nets (high-order feature detectors) without the dramatic slowdown we see in deep back-propagation networks. ? Cascade-Correlation is useful for incremental learning. in which new infonnation is added to an already-trained net. Once built. a feature detector is never cannibalized. It is available from that time on for producing outputs or more complex features. ? At any given time. we train only one layer of weights in the network. The rest of the network is constant. so results can be cached. ? There is never any need to propagate error signals backwards through network connections. A single residual error signal can be broadcast to all candidates. The weighted connections transmit signals in only one direction. eliminating one difference between these networks and biological synapses. ? The candidate units do not interact. except to pick a winner. Each candidate sees the same inputs and error signals. This limited communication makes the architecture attractive for parallel implementation. 4 RELATION TO OTHER WORK The principal differences between Cascade-Correlation and older learning architectures are the dynamic creation of hidden units. the way we stack the new units in multiple layers (with a fixed output layer). the freezing of units as we add them to the net. and the way we train new units by hill-climbing to maximize the unit's correlation with the residual error. The most interesting discovery is that by training one unit at a time instead of training the whole network at once. we can speed up the learning process considerably. while still creating a reasonably small net that generalizes well. A number of researchers [Ash. 1989.Moody. 1989] have investigated networks that add new units or receptive fields within a single layer in the course of learning. While single-layer systems are well-suited for some problems. these systems are incapable of creating higher-order feature detectors that combine the outputs of existing units. The idea of building feature detectors and then freezing them was inspired in part by the work of Waibel on modular networks [Waibel. 19891. but in his model the structure of the sub-networks must be fixed before learning begins. We know of only a few attempts to build up multi-layer networks as the learning progresses. Our decision to look at models in which each unit can see all pre-existing units was inspired to some extent by work on progressively deepening threshold-logic models by Merrick Furst and Jeff Jackson at Carnegie Mellon. (They are not actively pursuing this line at present.) Gallant [Gallant. 1986] briefly mentions a progressively deepening perceptron model (his "inverted pyramid model) in which units are frozen after being installed. However. he has concentrated most of his research effort on models in which new hidden units are generated at random rather than by a deliberate training process. The SONN model of Tenorio and Lee [Tenorio, 1989] builds a multiple-layer topology U 531 532 Fahlman and Lebiere to suit the problem at hand. Their algorithm places new -two-input units at randomly selected locations, using a simulated annealing search to keep only the most useful ones-a very different approach from ours. Acknowledgments We would like to thank Merrick Furst, Paul Gleichauf, and David Touretzlcy for asking good questions that helped to shape this work. This research was sponsored in part by the National Science Foundation (Contract EET-8716324) and in part by the Defense Advanced Research Projects Agency (Contract F3361S-87-C-1499). References [Ash, 1989] Ash, T. (1989) "Dynamic Node Creation in Back-Propagation Networks", Technical Report 8901, Institute for Cognitive Science, University of California, San Diego. [Fahlman, 1988] Fahlman, S. E. (1988) "Faster-Learning Variations on BackPropagation: An Empirical Study" in Proceedings of the 1988 Connectionist Models Summer School, Morgan Kaufmann. [Gallant, 1986] Gallant, S. I. (1986) "Three Constructive Algorithms for Network Learning" in Proceedings. 8th Annual Conference of the Cognitive Science Society. [Lang, 1988] Lang, K. J. and Witbrock, M. J. (1988) "Learning to Tell Two Spirals Apart" in Proceedings of the 1988 Connectionist Models Summer School, Morgan Kaufmann. [Moody, 1989] Moody, J. (1989) "Fast Learning in Multi-Resolution Hierarchies" in D. S. Touretzky (ed.), Advances in Neural Information Processing Systems 1, Morgan Kaufmann. [Rumelhart, 1986] Rumelhart, D. E., Hinton, G. E., and Williams, R. J. (1986) "Learning Internal Representations by Error Propagation" in Rumelhart, D. E. and McClelland, J. L.,Parallel Distributed Processing: Explorations in the Microstructure of Cognition, MIT Press. [Tenorio, 1989] Tenorio, M. E, and Lee, W. T. (1989) "Self-Organizing Neural Nets for the Identification Problem" in D. S. Touretzky (ed.), Advances in Neural Information Processing Systems 1, Morgan Kaufmann. [Tesauro, 1988] Tesauro, G. and Janssens, B. (1988) "Scaling Relations in BackPropagation Learning" in Complex Systems 2 39-44. [Waibel, 1989] Waibel, A. (1989) "Consonant Recognition by Modular Construction of Large Phonemic Time-Delay Neural Networks" in D. S. TouretzlcY (ed.), Advances in Neural Information Processing Systt ms 1, Morgan Kaufmann.
207 |@word trial:4 briefly:1 version:1 eliminating:1 simulation:2 propagate:3 quickprop:9 covariance:1 tried:1 pick:1 dramatic:1 mention:1 reduction:1 initial:1 score:1 ours:1 existing:6 comparing:2 activation:9 yet:4 lang:7 must:2 merrick:2 happen:1 shape:1 christian:1 asymptote:1 sponsored:1 update:1 progressively:2 half:3 leaf:1 fewer:1 guess:1 selected:1 short:1 farther:2 record:1 node:1 location:1 sigmoidal:2 five:2 direct:1 install:2 consists:1 combine:2 manner:1 roughly:1 multi:3 inspired:2 continuousvalued:1 automatically:2 actual:1 becomes:2 begin:5 classifies:2 provided:1 project:1 twisted:1 what:1 kind:1 developed:1 every:5 act:1 wrong:1 unit:101 uo:1 acceptably:1 producing:2 positive:2 before:2 installed:2 might:3 studied:1 examined:1 limited:1 range:1 averaged:3 acknowledgment:1 testing:1 backpropagation:2 nlogn:1 empirical:1 cascade:29 hyperbolic:1 got:1 pre:3 radial:1 suggest:1 get:2 cannot:1 operator:1 fonnula:1 interlocking:1 go:1 williams:1 resolution:1 immediately:1 rule:5 jackson:1 his:3 variation:1 transmit:1 target:1 suppose:1 diego:1 engage:2 hierarchy:1 construction:1 us:1 alexis:1 origin:1 pa:1 crossing:4 rumelhart:4 recognition:1 cut:1 observed:1 role:2 ep:2 connected:2 cycle:3 ran:3 environment:1 agency:1 complexity:2 touretzlcy:2 dynamic:2 ultimately:1 trained:11 creation:4 serve:1 creates:1 derivation:1 train:6 cacheing:1 fast:2 distinct:1 artificial:2 tell:1 whose:3 modular:2 larger:1 solve:7 final:2 obviously:1 differentiate:1 advantage:2 frozen:6 net:11 date:3 organizing:1 description:1 billion:1 convergence:1 produce:2 cached:2 incremental:1 converges:1 widrow:1 develop:2 measured:1 nearest:1 school:3 received:1 progress:2 phonemic:1 sonn:1 solves:1 direction:1 correct:2 exploration:1 settle:1 backprop:12 microstructure:1 generalization:1 biological:1 around:1 cognition:1 furst:2 purpose:1 currently:1 infonnation:1 create:2 weighted:3 rough:1 mit:1 gaussian:3 modified:2 rather:2 avoid:1 properly:2 greatly:1 eliminate:1 entire:3 lj:1 hidden:42 relation:2 expand:1 overall:1 classification:1 among:1 hoff:1 field:3 once:7 never:3 runtimes:1 look:1 cancel:1 plenty:1 report:3 connectionist:2 employ:1 few:2 randomly:1 simultaneously:1 national:1 individual:1 suit:1 attempt:4 decisively:1 multiply:1 mixture:2 accurate:1 partial:1 necessary:1 perfonnance:1 unless:1 divide:1 desired:1 re:1 minimal:1 recomputing:1 earlier:1 asking:1 retains:1 witbrock:5 delay:1 successful:1 reported:2 considerably:1 lee:2 contract:2 pool:6 together:1 quickly:2 moody:3 connectivity:1 again:3 satisfied:1 deepening:2 broadcast:1 external:1 creating:4 cognitive:2 derivative:2 li:1 actively:1 permanent:1 mime:1 helped:1 parallel:4 kaufmann:5 climbing:1 vp:1 identification:1 none:1 researcher:2 lumpy:1 detector:7 synapsis:1 touretzky:2 whenever:1 ed:3 lebiere:6 stop:2 experimenter:1 adjusting:2 popular:1 back:8 janssens:2 feed:1 higher:1 supervised:1 arranged:1 though:1 just:3 correlation:36 until:2 hand:1 receives:3 freezing:2 nonlinear:1 propagation:8 widespread:1 perhaps:1 believe:1 grows:2 building:2 requiring:1 symmetric:1 illustrated:1 attractive:1 during:3 self:1 m:1 trying:1 hill:1 motion:1 ranging:1 sigmoid:5 quarter:1 winner:1 million:2 analog:1 occurred:1 he:1 accumulate:1 mellon:2 significant:1 freeze:1 had:1 moving:2 add:5 own:1 dictated:1 apart:2 tesauro:4 slowness:2 certain:1 corp:1 incapable:1 binary:1 continue:1 inverted:1 morgan:5 greater:2 care:1 somewhat:2 preceding:1 eo:5 maximize:5 signal:7 smoother:1 multiple:2 reduces:1 technical:1 faster:4 offer:1 serial:2 inpu:1 controlled:1 essentially:1 histogram:1 sometimes:1 pyramid:1 receive:1 addition:2 annealing:1 diagram:1 median:1 rest:2 pass:1 ascent:1 call:1 backwards:1 spiral:7 affect:1 gleichauf:1 architecture:13 topology:3 reduce:1 idea:2 defense:1 effort:1 interpolates:1 repeatedly:4 deep:3 useful:4 ten:1 concentrated:1 mcclelland:1 wieland:2 deliberate:1 sign:3 delta:2 carnegie:2 key:1 threshold:1 changing:3 backward:2 wasted:1 sum:5 run:5 compete:1 powerful:1 place:1 family:1 almost:1 decide:1 pursuing:1 decision:1 patience:1 scaling:1 bit:1 layer:15 summer:2 fan:2 fold:2 annual:1 scanned:1 precisely:1 speed:2 extremely:1 attempting:1 speedup:5 developing:1 waibel:4 beneficial:1 smaller:2 cascadecorrelation:1 wi:2 turn:1 count:1 needed:2 know:1 unusual:1 available:2 generalizes:1 permanently:2 original:1 build:3 society:1 move:1 added:8 quantity:1 already:1 question:1 strategy:3 primary:1 receptive:2 gradient:1 thank:1 simulated:1 extent:1 useless:1 minimizing:1 negative:3 implementation:2 adjustable:1 perform:1 gallant:4 vertical:1 benchmark:4 beat:1 hinton:1 communication:1 precise:1 varied:3 stack:1 david:1 unpublished:1 required:4 trainable:1 extensive:1 connection:18 california:1 tremendous:1 able:1 below:1 pattern:10 scott:1 built:6 memory:1 residual:6 advanced:1 older:1 created:1 epoch:18 discovery:1 tangent:1 contributing:1 mixed:2 interesting:1 ash:3 foundation:1 lo:1 course:1 changed:1 repeat:1 fahlman:9 last:1 parity:5 slowdown:1 side:1 bias:1 perceptron:2 institute:1 neighbor:1 absolute:1 distributed:1 depth:2 world:1 preventing:1 stuck:1 made:1 forward:4 san:1 far:1 correlate:1 eet:1 logic:1 keep:1 active:7 incoming:5 pittsburgh:1 consonant:1 alternatively:1 search:1 reasonably:2 obtaining:1 improving:1 interact:2 boxed:1 investigated:2 complex:5 separator:1 mitre:1 whole:1 paul:1 positively:1 ff:1 sub:1 candidate:26 learns:3 formula:1 explored:3 adding:1 magnitude:3 suited:1 smoothly:1 tenorio:4 adjustment:1 determines:1 constantly:1 chance:1 goal:2 jeff:1 shortcut:1 change:2 hard:1 except:2 principal:1 pas:3 internal:1 constructive:1 dance:2
1,174
2,070
Latent Dirichlet Allocation David M. Blei, Andrew Y. Ng and Michael I. Jordan University of California, Berkeley Berkeley, CA 94720 Abstract We propose a generative model for text and other collections of discrete data that generalizes or improves on several previous models including naive Bayes/unigram, mixture of unigrams [6], and Hofmann's aspect model , also known as probabilistic latent semantic indexing (pLSI) [3]. In the context of text modeling, our model posits that each document is generated as a mixture of topics, where the continuous-valued mixture proportions are distributed as a latent Dirichlet random variable. Inference and learning are carried out efficiently via variational algorithms. We present empirical results on applications of this model to problems in text modeling, collaborative filtering, and text classification. 1 Introduction Recent years have seen the development and successful application of several latent factor models for discrete data. One notable example, Hofmann's pLSI/aspect model [3], has received the attention of many researchers, and applications have emerged in text modeling [3], collaborative filtering [7], and link analysis [1]. In the context of text modeling, pLSI is a "bag-of-words" model in that it ignores the ordering of the words in a document. It performs dimensionality reduction, relating each document to a position in low-dimensional "topic" space. In this sense, it is analogous to PCA, except that it is explicitly designed for and works on discrete data. A sometimes poorly-understood subtlety of pLSI is that, even though it is typically described as a generative model , its documents have no generative probabilistic semantics and are treated simply as a set of labels for the specific documents seen in the training set. Thus there is no natural way to pose questions such as "what is the probability of this previously unseen document?". Moreover, since each training document is treated as a separate entity, the pLSI model has a large number of parameters and heuristic "tempering" methods are needed to prevent overfitting. In this paper we describe a new model for collections of discrete data that provides full generative probabilistic semantics for documents. Documents are modeled via a hidden Dirichlet random variable that specifies a probability distribution on a latent, low-dimensional topic space. The distribution over words of an unseen document is a continuous mixture over document space and a discrete mixture over all possible topics. 2 2.1 Generative models for text Latent Dirichlet Allocation (LDA) model To simplify our discussion, we will use text modeling as a running example throughout this section, though it should be clear that the model is broadly applicable to general collections of discrete data. In LDA, we assume that there are k underlying latent topics according to which documents are generated, and that each topic is represented as a multinomial distribution over the IVI words in the vocabulary. A document is generated by sampling a mixture of these topics and then sampling words from that mixture. More precisely, a document of N words w = (W1,'" ,WN) is generated by the following process. First, B is sampled from a Dirichlet(a1,'" ,ak) distribution. This means that B lies in the (k - I)-dimensional simplex: Bi 2': 0, 2: i Bi = 1. Then, for each of the N words, a topic Zn E {I , ... , k} is sampled from a Mult(B) distribution p(zn = ilB) = Bi . Finally, each word Wn is sampled, conditioned on the znth topic, from the multinomial distribution p(wl zn). Intuitively, Bi can be thought of as the degree to which topic i is referred to in the document . Written out in full, the probability of a document is therefore the following mixture: p(w) = Ie (11 z~/(wnlzn; ,8)P( Zn IB?) p(B; a)dB, (1) where p(B ; a) is Dirichlet , p(znIB) is a multinomial parameterized by B, and p( Wn IZn;,8) is a multinomial over the words. This model is parameterized by the kdimensional Dirichlet parameters a = (a1,' .. ,ak) and a k x IVI matrix,8, which are parameters controlling the k multinomial distributions over words. The graphical model representation of LDA is shown in Figure 1. As Figure 1 makes clear, this model is not a simple Dirichlet-multinomial clustering model. In such a model the innermost plate would contain only W n ; the topic node would be sampled only once for each document; and the Dirichlet would be sampled only once for the whole collection. In LDA, the Dirichlet is sampled for each document, and the multinomial topic node is sampled repeatedly within the document. The Dirichlet is thus a component in the probability model rather than a prior distribution over the model parameters. We see from Eq. (1) that there is a second interpretation of LDA. Having sampled B, words are drawn iid from the multinomial/unigram model given by p(wIB) = 2::=1 p(wl z)p(z IB). Thus, LDA is a mixture model where the unigram models p(wIB) are the mixture components, and p(B ; a) gives the mixture weights. Note that unlike a traditional mixture of unigrams model, this distribution has an infinite o 1'0 '. Zn Wn Nd I D Figure 1: Graphical model representation of LDA. The boxes are plates representing replicates. The outer plate represents documents, while the inner plate represents the repeated choice of topics and words within a document. Figure 2: An example distribution on unigram models p(wIB) under LDA for three words and four topics. The triangle embedded in the x-y plane is the 2-D simplex over all possible multinomial distributions over three words. (E.g. , each of the vertices of the triangle corresponds to a deterministic distribution that assigns one of the words probability 1; the midpoint of an edge gives two of the words 0.5 probability each; and the centroid of the triangle is the uniform distribution over all 3 words). The four points marked with an x are the locations of the multinomial distributions p(wlz) for each of the four topics , and the surface shown on top of the simplex is an example of a resulting density over multinomial distributions given by LDA. number of continuously-varying mixture components indexed by B. The example in Figure 2 illustrates this interpretation of LDA as defining a random distribution over unigram models p(wIB). 2.2 Related models The mixture of unigrams model [6] posits that every document is generated by a single randomly chosen topic: (2) This model allows for different documents to come from different topics, but fails to capture the possibility that a document may express multiple topics. LDA captures this possibility, and does so with an increase in the parameter count of only one parameter: rather than having k - 1 free parameters for the multinomial p(z) over the k topics, we have k free parameters for the Dirichlet. A second related model is Hofmann's probabilistic latent semantic indexing (pLSI) [3], which posits that a document label d and a word ware conditionally independent given the hidden topic z : p(d, w) = L~=l p(wlz)p(zld)p(d). (3) This model does capture the possibility that a document may contain multiple topics since p(zld) serve as the mixture weights of the topics. However, a subtlety of pLSIand the crucial difference between it and LDA-is that d is a dummy index into the list of documents in the training set. Thus, d is a multinomial random variable with as many possible values as there are training documents, and the model learns the topic mixtures p(zld) only for those documents on which it is trained. For this reason, pLSI is not a fully generative model and there is no clean way to use it to assign probability to a previously unseen document. Furthermore, the number of parameters in pLSI is on the order of klVl + klDI, where IDI is the number of documents in the training set. Linear growth in the number of parameters with the size of the training set suggests that overfitting is likely to be a problem and indeed, in practice, a "tempering" heuristic is used to smooth the parameters of the model. 3 Inference and learning Let us begin our description of inference and learning problems for LDA by examining the contribution to the likelihood made by a single document. To simplify our notation, let w~ = 1 iff Wn is the jth word in the vocabulary and z~ = 1 iff Zn is the ith topic. Let j3ij denote p(w j = Ilzi = 1), and W = (WI, ... ,WN), Z = (ZI, ... ,ZN). Expanding Eq. (1), we have: (4) This is a hypergeometric function that is infeasible to compute exactly [4]. Large text collections require fast inference and learning algorithms and thus we have utilized a variational approach [5] to approximate the likelihood in Eq. (4). We use the following variational approximation to the log likelihood: logp(w; a, 13) log r :Ep(wlz; j3)p(zIB)p(B; a) qq~:,,Z", z:" ~~ dB le z > Eq[logp(wlz;j3) +logp(zIB) +logp(B;a) -logq(B,z; , ,?)], where we choose a fully factorized variational distribution q(B, z;" ?) q(B; ,) fIn q(Zn; ?n) parameterized by , and ?n, so that q(B; ,) is Dirichlet({), and q(zn; ?n) is MUlt(?n). Under this distribution, the terms in the variational lower bound are computable and differentiable, and we can maximize the bound with respect to, and ? to obtain the best approximation to p(w;a,j3). Note that the third and fourth terms in the variational bound are not straightforward to compute since they involve the entropy of a Dirichlet distribution, a (k - I)-dimensional integral over B which is expensive to compute numerically. In the full version of this paper, we present a sequence of reductions on these terms which use the log r function and its derivatives. This allows us to compute the integral using well-known numerical routines. Variational inference is coordinate ascent in the bound on the probability of a single document. In particular, we alternate between the following two equations until the objective converges: (5) ,i + 2:~=1 ?ni derivative of the log r function. ai (6) where \]i is the first Note that the resulting variational parameters can also be used and interpreted as an approximation of the parameters of the true posterior. In the current paper we focus on maximum likelihood methods for parameter estimation. Given a collection of documents V = {WI' ... ' WM}, we utilize the EM algorithm with a variational E step, maximizing a lower bound on the log likelihood: M logp(V) 2:: l:= Eqm [logp(B, z, w)]- Eqm [logqm(B, z)]. (7) m=l The E step refits qm for each document by running the inference step described above. The M step optimizes Eq. (7) with respect to the model parameters a and (3. For the multinomial parameters (3ij we have the following M step update equation: M (3ij ex: Iwml l:= l:= ?>mniwtnn? (8) m=l n=l The Dirichlet parameters ai are not independent of each other and we apply Newton-Raphson to optimize them: The variational EM algorithm alternates between maximizing Eq. (7) with respect to qm and with respect to (a, (3) until convergence. 4 Experiments and Examples We first tested LDA on two text corpora. 1 The first was drawn from the TREC AP corpus, and consisted of 2500 news articles, with a vocabulary size of IVI = 37,871 words. The second was the CRAN corpus, consisting of 1400 technical abstracts, with IVI = 7747 words. We begin with an example showing how LDA can capture multiple-topic phenomena in documents. By examining the (variational) posterior distribution on the topic mixture q(B; ')'), we can identify the topics which were most likely to have contributed to many words in a given document; specifically, these are the topics i with the largest ')'i. Examining the most likely words in the corresponding multinomials can then further tell us what these topics might be about. The following is an article from the TREC collection. The William R andolph Hearst Foundation will give $1.25 million to Lincoln Center, Metropolitan Opera Co., New York Philharmonic and Juilliard School. "Our board felt that we had a real opportunity to make a mark on the future of the performing arts with these grants an act every bit as important as our traditional areas of support in health , medical research, education and the social services," Hearst Foundation President Randolph A. Hearst said Monday in announcing the grants. Lincoln Center's share will be $200,000 for its new building, which will house young artists and provide new public facilities. The Metropolitan Opera Co. and New York Philharmonic will receive $400 ,000 each. The Juilliard School, where music and the performing arts are taught, will get $250,000 . The Hearst Foundation, a leading supporter of the Lincoln Center Consolidated Corporate Fund, will make its usual annual $100,000 donation, too. Figure 3 shows the Dirichlet parameters of the corresponding variational distribution for those topics where ')'i > 1 (k = 100) , and also lists the top 15 words (in iTo enable repeated large scale comparison of various models on large corpora, we implemented our variational inference algorithm on a parallel computing cluster. The (bottleneck) E step is distributed across nodes so that the qm for different documents are calculated in parallel. I" Topic 1 Topic 2 Topic 3 Topic 4 Topic 5 SCHOOL SAID STUDENTS BOARD SCHOOLS STUDENT TEACHER POLICE PROGRAM TEACHERS MEMBERS YEAROLD GANG DEPARTMENT MILLION YEAR SAID SALES BILLION TOTAL SHARE EARNINGS PROFIT QUARTER ORDERS LAST DEC REVENUE SAID AIDS HEALTH DISEASE VIRUS CHILDREN BLOOD PATIENTS TREATMENT STUDY IMMUNE CANCER PEOPLE PERCENT SAID NEW PRESIDENT CHIEF CHAIRMAN EXECUTIVE VICE YEARS COMPANY YORK SCHOOL TWO TODAY COLUMBIA SAID NEW MUSIC YEAR THEATER MUSICAL BAND PLAY WON TWO AVAILABLE AWARD OPERA BEST Figure 3: The Dirichlet parameters where Ii > 1 (k = 100), and the top 15 words from the corresponding topics, for the document discussed in the text. __ LDA -x- pLSI .. ~ pLSI(00 lemper) MIx1Un;grams v ? ram woo ' .~ 4500 ',.. _ .l\ ! -><--------------- k (number of topics) k (number of topiCS) Figure 4: Perplexity results on the CRAN and AP corpora for LDA, pLSI, mixture of unigrams, and t he unigram model. order) from these topics. This document is mostly a combination of words about school policy (topic 4) and music (topic 5). The less prominent topics reflect other words about education (topic 1) , finance (topic 2), and health (topic 3). 4.1 Formal evaluation: Perplexity To compare the generalization performance of LDA with other models, we computed the perplexity of a test set for the AP and CRAN corpora. The perplexity, used by convention in language modeling, is monotonically decreasing in the likelihood of the test data, and can be thought of as the inverse of the per-word likelihood. More formally, for a test set of M documents, perplexity(Vtest ) = exp (-l:m logp(wm)/ l:m Iwml}. We compared LDA to both the mixture of unigrams and pLSI described in Section 2.2. We trained the pLSI model with and without tempering to reduce overfitting. When tempering, we used part of the test set as the hold-out data, thereby giving it a slight unfair advantage. As mentioned previously, pLSI does not readily generate or assign probabilities to previously unseen documents; in our experiments, we assigned probability to a new document d by marginalizing out the dummy training set indices 2 : pew ) = l: d( rr : =1l:z p(w n lz)p(z ld))p(d) . 2 A second natural method, marginalizing out d and z to form a unigram model using the resulting p(w)'s, did not perform well (its performance was similar to the standard unigram model). 1-:- ~Dc.~UrUg,ams I W' ? M" ~ x NaiveBaes k (number of topics) k (number of topics) Figure 5: Results for classification (left) and collaborative filtering (right) Figure 4 shows the perplexity for each model and both corpora for different values of k. The latent variable models generally do better than the simple unigram model. The pLSI model severely overfits when not tempered (the values beyond k = 10 are off the graph) but manages to outperform mixture of unigrams when tempered. LDA consistently does better than the other models. To our knowledge, these are by far the best text perplexity results obtained by a bag-of-words model. 4.2 Classification We also tested LDA on a text classification task. For each class c, we learn a separate model p(wlc) of the documents in that class. An unseen document is classified by picking argmaxcp(Clw) = argmaxcp(wlc)p(c). Note that using a simple unigram distribution for p(wlc) recovers the traditional naive Bayes classification model. Using the same (standard) subset of the WebKB dataset as used in [6], we obtained classification error rates illustrated in Figure 5 (left). In all cases, the difference between LDA and the other algorithms' performance is statistically significant (p < 0.05). 4.3 Collaborative filtering Our final experiment utilized the EachMovie collaborative filtering dataset. In this dataset a collection of users indicates their preferred movie choices. A user and the movies he chose are analogous to a document and the words in the document (respectively) . The collaborative filtering task is as follows. We train the model on a fully observed set of users. Then, for each test user, we are shown all but one of the movies that she liked and are asked to predict what the held-out movie is. The different algorithms are evaluated according to the likelihood they assign to the held-out movie. More precisely define the predictive perplexity on M test users to be exp( - ~~=llogP(WmNd lwml' ... ,Wm(Nd-l))/M) . With 5000 training users, 3500 testing users, and a vocabulary of 1600 movies, we find predictive perplexities illustrated in Figure 5 (right). 5 Conclusions We have presented a generative probabilistic framework for modeling the topical structure of documents and other collections of discrete data. Topics are represented explicitly via a multinomial variable Zn that is repeatedly selected, once for each word, in a given document. In this sense, the model generates an allocation of the words in a document to topics. When computing the probability of a new document, this unknown allocation induces a mixture distribution across the words in the vocabulary. There is a many-to-many relationship between topics and words as well as a many-to-many relationship between documents and topics. While Dirichlet distributions are often used as conjugate priors for multinomials in Bayesian modeling, it is preferable to instead think of the Dirichlet in our model as a component of the likelihood. The Dirichlet random variable e is a latent variable that gives generative probabilistic semantics to the notion of a "document" in the sense that it allows us to put a distribution on the space of possible documents. The words that are actually obtained are viewed as a continuous mixture over this space, as well as being a discrete mixture over topics. 3 The generative nature of LDA makes it easy to use as a module in more complex architectures and to extend it in various directions. We have already seen that collections of LDA can be used in a classification setting. If the classification variable is treated as a latent variable we obtain a mixture of LDA models, a useful model for situations in which documents cluster not only according to their topic overlap, but along other dimensions as well. Another extension arises from generalizing LDA to consider Dirichlet/multinomial mixtures of bigram or trigram models, rather than the simple unigram models that we have considered here. Finally, we can readily fuse LDA models which have different vocabularies (e.g., words and images); these models interact via a common abstract topic variable and can elegantly use both vocabularies in determining the topic mixture of a given document. Acknowledgments A. Ng is supported by a Microsoft Research fellowship. This work was also supported by a grant from Intel Corporation, NSF grant IIS-9988642, and ONR MURI N00014-00-1-0637. References [1] D. Cohn and T. Hofmann. The missing link- A probabilistic model of document content and hypertext connectivity. In Advances in Neural Information Processing Systems 13, 2001. [2] P.J. Green and S. Richardson. Modelling heterogeneity with and without the Dirichlet process. Technical Report, University of Bristol, 1998. [3] T. Hofmann. Probabilistic latent semantic indexing. Proceedings of th e Twenty-Second Annual International SIGIR Conference, 1999. [4] T. J. Jiang, J. B. Kadane, and J. M. Dickey. Computation of Carlson's multiple hypergeometric functions r for Bayesian applications. Journal of Computational and Graphical Statistics, 1:231- 251 , 1992. [5] M. I. Jordan , Z. Ghahramani, T. S. Jaakkola, and L. K. Saul. Introduction to variational methods for graphical models. Machine Learning, 37:183- 233, 1999. [6] K. Nigam, A. Mccallum, S. Thrun, and T. Mitchell. Text classification from labeled and unlabeled documents using EM. Machine Learning, 39(2/3):103- 134, 2000. [7] A. Popescul, L. H. Ungar, D. M. Pennock, and S. Lawrence. Probabilistic models for unified collaborative and content-based recommendation in sparse-data environments. In Uncertainty in Artificial Intelligence, Proceedings of the Seventeenth Conference, 2001. 3These remarks also distinguish our model from the Bayesian Dirichlet/Multinomial allocation model (DMA)of [2], which is a finite alternative to the Dirichlet process . The DMA places a mixture of Dirichlet priors on p(wl z ) and sets O i = 00 for all i .
2070 |@word version:1 bigram:1 proportion:1 nd:2 innermost:1 profit:1 thereby:1 ld:1 reduction:2 document:58 current:1 virus:1 written:1 readily:2 numerical:1 hofmann:5 designed:1 update:1 fund:1 generative:9 selected:1 intelligence:1 plane:1 mccallum:1 ith:1 randolph:1 blei:1 provides:1 node:3 location:1 idi:1 monday:1 earnings:1 along:1 indeed:1 decreasing:1 company:1 begin:2 webkb:1 moreover:1 underlying:1 notation:1 factorized:1 what:3 interpreted:1 consolidated:1 unified:1 corporation:1 berkeley:2 every:2 act:1 growth:1 finance:1 exactly:1 preferable:1 qm:3 sale:1 grant:4 medical:1 service:1 understood:1 severely:1 ak:2 jiang:1 ware:1 ap:3 might:1 chose:1 suggests:1 co:2 bi:4 statistically:1 seventeenth:1 acknowledgment:1 testing:1 practice:1 area:1 empirical:1 mult:2 thought:2 word:36 get:1 unlabeled:1 put:1 context:2 optimize:1 deterministic:1 center:3 maximizing:2 missing:1 straightforward:1 attention:1 sigir:1 assigns:1 theater:1 notion:1 coordinate:1 analogous:2 qq:1 president:2 controlling:1 today:1 play:1 user:7 expensive:1 utilized:2 muri:1 logq:1 labeled:1 ep:1 observed:1 module:1 capture:4 hypertext:1 news:1 ordering:1 disease:1 mentioned:1 environment:1 asked:1 trained:2 predictive:2 serve:1 triangle:3 represented:2 various:2 llogp:1 train:1 fast:1 describe:1 artificial:1 tell:1 emerged:1 heuristic:2 valued:1 statistic:1 unseen:5 richardson:1 think:1 final:1 sequence:1 differentiable:1 advantage:1 rr:1 propose:1 iff:2 poorly:1 lincoln:3 description:1 billion:1 convergence:1 cluster:2 liked:1 converges:1 donation:1 andrew:1 pose:1 ij:2 school:6 received:1 eq:6 implemented:1 come:1 convention:1 direction:1 posit:3 announcing:1 enable:1 public:1 education:2 require:1 wlc:3 assign:3 ungar:1 generalization:1 extension:1 hold:1 considered:1 exp:2 lawrence:1 predict:1 trigram:1 estimation:1 applicable:1 bag:2 label:2 largest:1 wl:3 vice:1 metropolitan:2 zib:2 rather:3 varying:1 jaakkola:1 focus:1 she:1 consistently:1 modelling:1 likelihood:9 indicates:1 centroid:1 sense:3 am:1 inference:7 typically:1 hidden:2 semantics:3 classification:9 development:1 art:2 once:3 having:2 ng:2 sampling:2 represents:2 future:1 simplex:3 report:1 ilb:1 simplify:2 randomly:1 consisting:1 william:1 microsoft:1 possibility:3 evaluation:1 replicates:1 mixture:27 held:2 dickey:1 edge:1 integral:2 indexed:1 modeling:8 zn:10 logp:7 vertex:1 subset:1 uniform:1 successful:1 examining:3 too:1 kadane:1 teacher:2 density:1 international:1 ie:1 probabilistic:9 off:1 picking:1 kdimensional:1 michael:1 continuously:1 w1:1 connectivity:1 reflect:1 choose:1 derivative:2 leading:1 student:2 notable:1 explicitly:2 unigrams:6 overfits:1 wm:3 bayes:2 parallel:2 collaborative:7 contribution:1 opera:3 ni:1 musical:1 efficiently:1 identify:1 bayesian:3 artist:1 manages:1 iid:1 researcher:1 bristol:1 classified:1 recovers:1 sampled:8 dataset:3 treatment:1 mitchell:1 knowledge:1 improves:1 dimensionality:1 routine:1 actually:1 evaluated:1 though:2 box:1 furthermore:1 until:2 cran:3 cohn:1 lda:27 building:1 contain:2 true:1 wlz:4 consisted:1 facility:1 assigned:1 semantic:3 illustrated:2 conditionally:1 won:1 prominent:1 plate:4 performs:1 percent:1 image:1 variational:14 common:1 multinomial:19 quarter:1 million:2 extend:1 interpretation:2 discussed:1 relating:1 numerically:1 he:2 slight:1 significant:1 ai:2 pew:1 language:1 had:1 immune:1 vtest:1 surface:1 posterior:2 plsi:15 recent:1 optimizes:1 perplexity:9 n00014:1 onr:1 wib:4 tempered:2 seen:3 maximize:1 eqm:2 monotonically:1 ii:2 full:3 multiple:4 corporate:1 eachmovie:1 smooth:1 technical:2 raphson:1 award:1 a1:2 j3:3 patient:1 sometimes:1 dec:1 receive:1 fellowship:1 crucial:1 unlike:1 ivi:4 ascent:1 pennock:1 db:2 member:1 jordan:2 easy:1 wn:6 zi:1 architecture:1 inner:1 reduce:1 computable:1 supporter:1 bottleneck:1 pca:1 york:3 repeatedly:2 remark:1 generally:1 useful:1 clear:2 involve:1 band:1 induces:1 generate:1 specifies:1 outperform:1 nsf:1 dummy:2 per:1 broadly:1 discrete:8 express:1 taught:1 four:3 blood:1 tempering:4 drawn:2 prevent:1 clean:1 utilize:1 ram:1 graph:1 fuse:1 clw:1 year:4 inverse:1 parameterized:3 fourth:1 uncertainty:1 place:1 throughout:1 bit:1 bound:5 distinguish:1 annual:2 gang:1 precisely:2 felt:1 generates:1 aspect:2 performing:2 department:1 according:3 alternate:2 combination:1 conjugate:1 across:2 em:3 wi:2 intuitively:1 indexing:3 equation:2 previously:4 count:1 needed:1 generalizes:1 available:1 apply:1 alternative:1 top:3 clustering:1 dirichlet:25 running:2 graphical:4 opportunity:1 newton:1 music:3 carlson:1 giving:1 ghahramani:1 objective:1 question:1 already:1 chairman:1 usual:1 traditional:3 said:6 link:2 separate:2 thrun:1 entity:1 outer:1 topic:55 reason:1 modeled:1 index:2 relationship:2 izn:1 mostly:1 popescul:1 refit:1 policy:1 contributed:1 perform:1 unknown:1 twenty:1 fin:1 finite:1 defining:1 situation:1 heterogeneity:1 trec:2 dc:1 topical:1 hearst:4 police:1 david:1 california:1 hypergeometric:2 zld:3 beyond:1 program:1 including:1 green:1 overlap:1 treated:3 natural:2 representing:1 movie:6 carried:1 woo:1 naive:2 health:3 columbia:1 text:14 prior:3 marginalizing:2 determining:1 embedded:1 fully:3 allocation:5 filtering:6 revenue:1 foundation:3 executive:1 degree:1 article:2 share:2 cancer:1 supported:2 last:1 free:2 jth:1 infeasible:1 formal:1 saul:1 midpoint:1 sparse:1 distributed:2 calculated:1 vocabulary:7 gram:1 dimension:1 ignores:1 collection:10 made:1 lz:1 far:1 social:1 approximate:1 preferred:1 overfitting:3 corpus:7 continuous:3 latent:12 chief:1 learn:1 nature:1 ca:1 expanding:1 nigam:1 interact:1 complex:1 elegantly:1 did:1 whole:1 dma:2 repeated:2 child:1 referred:1 intel:1 board:2 aid:1 fails:1 position:1 lie:1 house:1 unfair:1 ib:2 third:1 ito:1 learns:1 young:1 unigram:11 specific:1 showing:1 list:2 conditioned:1 illustrates:1 entropy:1 generalizing:1 simply:1 likely:3 subtlety:2 recommendation:1 corresponds:1 marked:1 viewed:1 content:2 infinite:1 except:1 specifically:1 total:1 formally:1 mark:1 support:1 people:1 arises:1 tested:2 phenomenon:1 ex:1
1,175
2,071
Distribution of Mutual Information Marcus Hutter IDSIA, Galleria 2, CH-6928 Manno-Lugano, Switzerland marcus@idsia.ch http://www.idsia.ch/- marcus Abstract The mutual information of two random variables z and J with joint probabilities {7rij} is commonly used in learning Bayesian nets as well as in many other fields. The chances 7rij are usually estimated by the empirical sampling frequency nij In leading to a point estimate J(nij In) for the mutual information. To answer questions like "is J (nij In) consistent with zero?" or "what is the probability that the true mutual information is much larger than the point estimate?" one has to go beyond the point estimate. In the Bayesian framework one can answer these questions by utilizing a (second order) prior distribution p( 7r) comprising prior information about 7r. From the prior p(7r) one can compute the posterior p(7rln), from which the distribution p(Iln) of the mutual information can be calculated. We derive reliable and quickly computable approximations for p(Iln). We concentrate on the mean, variance, skewness, and kurtosis , and non-informative priors. For the mean we also give an exact expression. Numerical issues and the range of validity are discussed. 1 Introduction The mutual information J (also called cross entropy) is a widely used information theoretic measure for the stochastic dependency of random variables [CT91, SooOO] . It is used, for instance, in learning Bayesian nets [Bun96, Hec98] , where stochastically dependent nodes shall be connected. The mutual information defined in (1) can be computed if the joint probabilities {7rij} of the two random variables z and J are known. The standard procedure in the common case of unknown chances 7rij is to use the sample frequency estimates n~; instead, as if they were precisely known probabilities; but this is not always appropriate. Furthermore, the point estimate J (n~; ) gives no clue about the reliability of the value if the sample size n is finite. For instance, for independent z and J, J(7r) =0 but J(n~;) = O(n- 1 / 2 ) due to noise in the data. The criterion for judging dependency is how many standard deviations J(":,;) is away from zero. In [KJ96, Kle99] the probability that the true J(7r) is greater than a given threshold has been used to construct Bayesian nets. In the Bayesian framework one can answer these questions by utilizing a (second order) prior distribution p(7r),which takes account of any impreciseness about 7r. From the prior p(7r) one can compute the posterior p(7rln), from which the distribution p(Iln) of the mutual information can be obtained. The objective of this work is to derive reliable and quickly computable analytical expressions for p(1ln). Section 2 introduces the mutual information distribution, Section 3 discusses some results in advance before delving into the derivation. Since the central limit theorem ensures that p(1ln) converges to a Gaussian distribution a good starting point is to compute the mean and variance of p(1ln). In section 4 we relate the mean and variance to the covariance structure of p(7rln). Most non-informative priors lead to a Dirichlet posterior. An exact expression for the mean (Section 6) and approximate expressions for t he variance (Sections 5) are given for the Dirichlet distribution. More accurate estimates of the variance and higher central moments are derived in Section 7, which lead to good approximations of p(1ln) even for small sample sizes. We show that the expressions obtained in [KJ96, Kle99] by heuristic numerical methods are incorrect. Numerical issues and the range of validity are briefly discussed in section 8. 2 Mutual Information Distribution We consider discrete random variables Z E {l, ... ,r} and J E {l, ... ,s} and an i.i.d. random process with samples (i,j) E {l ,... ,r} x {l, ... ,s} drawn with joint probability 7rij. An important measure of the stochastic dependence of z and J is the mutual information T 1(7r) = S 7rij L L 7rij log ~ = L 7rij log 7rij - L 7ri+ log7ri+ - L 7r +j log7r i=1 j = 1 H +J ij i +j' (1) j log denotes the natural logarithm and 7ri+ = Lj7rij and 7r +j = L i 7rij are marginal probabilities. Often one does not know the probabilities 7rij exactly, but one has a sample set with nij outcomes of pair (i,j). The frequency irij := n~j may be used as a first estimate of the unknown probabilities. n:= L ijnij is the total sample size. This leads to a point (frequency) estimate 1(ir) = Lij n~j logn:~:j for the mutual informat ion (per sample). Unfortunately the point estimation 1(ir) gives no information about its accuracy. In the Bayesian approach to this problem one assumes a prior (second order) probability density p( 7r) for the unknown probabilities 7rij on the probability simplex. From this one can compute the posterior distribution p( 7rln) cxp( 7r) rr ij7r~;j (the nij are multinomially distributed). This allows to compute the posterior probability density of the mutual information.1 p(Iln) = 2The 80 f 8(1(7r) - I)p(7rln)d TS 7r (2) distribution restricts the integral to 7r for which 1(7r) =1. For large sam- 1 I(7r) denotes the mutual information for the specific chances 7r, whereas I in the context above is just some non-negative real number. I will also denote the mutual information random variable in the expectation E [I] and variance Var[I]. Expectaions are always w.r.t. to the posterior distribution p(7rln). 2Since O~I(7r) ~Imax with sharp upper bound Imax :=min{logr,logs}, the integral may be restricted to which shows that the domain of p(Iln) is [O,Imax] . J:mam, pIe size n ---+ 00, p(7rln) is strongly peaked around 7r = it and p(Iln) gets strongly peaked around the frequency estimate I = I(it). The mean E[I] = fooo Ip(Iln) dI = f I(7r)p(7rln)dTs 7r and the variance Var[I] =E[(I - E[I])2] = E[I2]- E[Ij2 are of central interest. 3 Results for I under the Dirichlet P (oste )rior Most 3 non-informative priors for p(7r) lead to a Dirichlet posterior distribution nij -1 h p (7r I) n ex: IT ij 7rij WI?th?IIIt erpre t a t?IOn nij - n, + n,ij , were n 'ij are th e numb er ij , of samples (i,j), and n~j comprises prior information (1 for the uniform prior, ~ for Jeffreys' prior, 0 for Haldane's prior, -?:s for Perks' prior [GCSR95]). In principle this allows to compute the posterior density p(Iln) of the mutual information. In sections 4 and 5 we expand the mean and variance in terms of n- 1 : E[I] ~ nij I nijn L...J og - - .. n ni+n+j 'J + (r - 1)(8 - 1) 2n + O( -2) n , (3) Var[I] The first term for the mean is just the point estimate I(it). The second term is a small correction if n ? r? 8. Kleiter [KJ96, Kle99] determined the correction by Monte Carlo studies as min {T2~1 , 8;;;,1 }. This is wrong unless 8 or rare 2. The expression 2E[I]/n they determined for the variance has a completely different structure than ours. Note that the mean is lower bounded by co~st. +O(n- 2 ), which is strictly positive for large, but finite sample sizes, even if z and J are statistically independent and independence is perfectly represented in the data (I (it) = 0). On the other hand, in this case, the standard deviation u= y'Var(I) '" ~ ",E[I] correctly indicates that the mean is still consistent with zero. Our approximations (3) for the mean and variance are good if T~8 is small. The central limit theorem ensures that p(Iln) converges to a Gaussian distribution with mean E[I] and variance Var[I]. Since I is non-negative it is more appropriate to approximate p(II7r) as a Gamma (= scaled X2 ) or log-normal distribution with mean E[I] and variance Var[I], which is of course also asymptotically correct. A systematic expansion in n -1 of the mean, variance, and higher moments is possible but gets arbitrarily cumbersome. The O(n - 2) terms for the variance and leading order terms for the skewness and kurtosis are given in Section 7. For the mean it is possible to give an exact expression 1 E[I] = - L nij[1jJ(nij + 1) -1jJ(ni+ + 1) -1jJ(n+j + 1) + 1jJ(n + 1)] n .. (4) 'J with 1jJ(n+1)=-,),+L~= lt=logn+O(~) for integer n. See Section 6 for details and more general expressions for 1jJ for non-integer arguments. There may be other prior information available which cannot be comprised in a Dirichlet distribution. In this general case, the mean and variance of I can still be 3But not all priors which one can argue to be non-informative lead to Dirichlet posteriors. Brand [Bragg] (and others), for instance, advocate the entropic prior p( 7r) ex e-H(rr). related to the covariance structure of p(7fln), which will be done in the following Section. 4 Approximation of Expectation and Variance of I In the following let fr ij := E[7fij]. Since p( 7fln) is strongly peaked around 7f = fr for large n we may expand J(7f) around fr in the integrals for the mean and the variance. With I:::..ij :=7fij -frij and using L: ij7fij = 1 = L:ijfrij we get for the expansion of (1) fr .. ) 1:::.. 2 . 1:::.. 2 1:::.. 2 . J(7f) = J(fr) + 2)og ( ~ I:::..ij + ----}J-~~+O(1:::..3). (5) .. 7fi+7f+j .. 27fij . 27fi+ . 27f+j 2J 2J 2 J L L L Taking the expectation, the linear term E[ I:::.. ij ] = a drops out. The quadratic terms E[ I:::..ij I:::..kd = Cov( 7fij ,7fkl) are the covariance of 7f under distribution p(7fln) and are proportional to n- 1 . It can be shown that E[1:::..3] ,,-,n- 2 (see Section 7). 1 " , (bikbjl bik - -Abjl) COV7fij,7fkl ( ) +On (-2) . ( A) +-~ [ ] = J7f EJ - A - - -A2 ijkl 7fij 7fi+ 7f+j The Kronecker delta bij is 1 for i = j and order in n - 1 is a otherwise. (6) The variance of J in leading (7) where :t means = up to terms of order n -2. So the leading order variance and the leading and next to leading order mean of the mutual information J(7f) can be expressed in terms of the covariance of 7f under the posterior distribution p(7fln). 5 The Second Order Dirichlet Distribution Noninformative priors for p(7f) are commonly used if no additional prior information is available. Many non-informative choices (uniform, Jeffreys' , Haldane's, Perks', prior) lead to a Dirichlet posterior distribution: II 1 n;j - 1 ( N(n) .. 7fij b 7f++ - 1) with normalization 2J N(n) (8) where r is the Gamma function, and nij = n~j + n~j, where n~j are the number of samples (i,j), and n~j comprises prior information (1 for the uniform prior, ~ for Jeffreys' prior, a for Haldane's prior, -!s for Perks' prior) . Mean and covariance of p(7fln) are [ ] =nij 7fij:= E7fij -, A n (9) Inserting this into (6) and (7) we get after some algebra for the mean and variance of the mutual information I(7r) up to terms of order n- 2 : E[I] J + (r - 1)(8 - 1) 2(n + 1) ~1 (K - J2) + Var[I] n+ + O( -2) n , (10) (11) 0(n-2), (12) (13) J and K (and L, M, P, Q defined later) depend on 7rij = ":,j only, i.e. are 0(1) in n. Strictly speaking we should expand n~l = ~+0(n-2), i.e. drop the +1, but the exact expression (9) for the covariance suggests to keep the +1. We compared both versions with the exact values (from Monte-Carlo simulations) for various parameters 7r. In most cases the expansion in n~l was more accurate, so we suggest to use this variant. 6 Exact Value for E[I] It is possible to get an exact expression for the mean mutual information E[I] under the Dirichlet distribution. By noting that xlogx= d~x,6I,6= l' (x = {7rij,7ri+ ,7r+j}), one can replace the logarithms in the last expression of (1) by powers. From (8) we see that E[ (7rij ),6] = ~i~:~ t~~~;l. Taking the derivative and setting ,8 = 1 we get E[7rij log 7rij] d 1 = d,8E[(7rij) ,6],6=l = ;;: 2:::: nij[1j!(nij + 1) -1j!(n + 1)]. "J The 1j! function has the following properties (see [AS74] for details) 1j!(z) = dlogf(z) dz = n- l 1j!(n) = -"( + L f'(z) f(z)' 1 k' 1j!(z + 1) = log z + 1 1 1 2z - 12z2 + O( Z4)' 1j!(n +~) = -"( + 2log2 + 2 k=l n L 1 2k _ l' (14) k=l The value of the Euler constant "( is irrelevant here, since it cancels out. Since the marginal distributions of 7ri+ and 7r+j are also Dirichlet (with parameters ni+ and n+j) we get similarly -n1 L. n+j[1j!(n+j + 1) -1j!(n + 1)]. J Inserting this into (1) and rearranging terms we get the exact expression 4 E[I] 1 =- L nij[1j! (nij + 1) -1j!(ni+ + 1) -1j!(n+j + 1) + 1j!(n + 1)] n .. 4This expression has independently been derived in [WW93]. (15) For large sample sizes, 'Ij;(z + 1) ~ logz and (15) approaches the frequency estimate I(7r) as it should be. Inserting the expansion 'Ij;(z + 1) = logz + + ... into (15) we also get the correction term (r - 11~s - 1) of (3). 2\ The presented method (with some refinements) may also be used to determine an exact expression for the variance of I(7f). All but one term can be expressed in terms of Gamma functions. The final result after differentiating w.r.t. (31 and (32 can be represented in terms of 'Ij; and its derivative 'Ij;' . The mixed term E[( 7fi+ )131 (7f +j )132] is more complicated and involves confluent hypergeometric functions, which limits its practical use [WW93] . 7 Generalizations A systematic expansion of all moments of p(Iln) to arbitrary order in n -1 is possible, but gets soon quite cumbersome. For the mean we already gave an exact expression (15), so we concentrate here on the variance, skewness and the kurtosis of p(Iln). The 3rd and 4th central moments of 7f under the Dirichlet distribution are ( )2( ) [27ra7rb7rc - 7ra7rbc5bc - 7rb7rcc5ca - 7rc7rac5ab n+l n+2 + 7rac5abc5bc] (16) ~2 [37ra7rb7rc7rd - jrc!!d!!a c5 ab - A7rbjrdA7rac5ac - A7rbA7rcA7rac5ad -7fa7fd7fbc5bc - 7fa7fc7fbc5bd - 7fa7fb7fcc5cd (17) +7ra7rcc5abc5cd + 7ra7rbc5acc5bd + 7ra7rbc5adc5bc] + O(n- 3) with a=ij, b= kl, ... E {1, ... ,r} x {1, ... ,8} being double indices, c5 ab =c5ik c5 jl ,... 7rij = n~j ? Expanding D.. k = (7f_7r)k in E[D..aD..b ... ] leads to expressions containing E[7fa7fb ... ], which can be computed by a case analysis of all combinations of equal/unequal indices a,b,c, ... using (8). Many terms cancel leading to the above expressions. They allow to compute the order n- 2 term of the variance of I(7f). Again, inspection of (16) suggests to expand in [(n+l)(n+2)]-1, rather than in n- 2 . The variance in leading and next to leading order is Var[I] M K - J2 + M + (r - 1)(8 - 1)(~ - J) - Q + O(n - 3) (n + l)(n + 2) n+ 1 (18) L (19) ij (~- _1_ _ _ 1_ nij ni+ n+j +~) nij log nijn , n ni+n+j 2 l-L~? Q ij ni+n+j (20) J and K are defined in (12) and (13). Note that the first term ~+f also contains second order terms when expanded in n -1. The leading order terms for the 3rd and 4th central moments of p(Iln) are L .- '""" nij ~- j I og--nij n n 32 [K - J 2 n ni+n+j F+ O(n - 3 ), from which the skewness and kurtosis can be obtained by dividing by Var[Ij3/2 and Var[IF respectively. One can see that the skewness is of order n- 1 / 2 and the kurtosis is 3 + 0 (n - 1). Significant deviation of the skewness from a or the kurtosis from 3 would indicate a non-Gaussian I. They can be used to get an improved approximation for p(Iln) by making, for instance, an ansatz and fitting the parameters b, c, jJ" and (j-2 to the mean, variance, skewness, and kurtosis expressions above. Po is the Normal or Gamma distribution (or any other distribution with Gaussian limit). From this, quantiles p(I>I*ln):= fI:'p(Iln) dI, needed in [KJ96, Kle99], can be computed. A systematic expansion of arbitrarily high moments to arbitrarily high order in n- 1 leads, in principle, to arbitrarily accurate estimates. 8 Numerics There are short and fast implementations of'if;. The code of the Gamma function in [PFTV92], for instance, can be modified to compute the 'if; function. For integer and half-integer values one may create a lookup table from (14) . The needed quantities J, K, L, M, and Q (depending on n) involve a double sum, P only a single sum, and the r+s quantities J i + and J+ j also only a single sum. Hence, the computation time for the (central) moments is of the same order O(r?s) as for the point estimate (1). "Exact" values have been obtained for representative choices of 7rij, r, s, and n by Monte Carlo simulation. The 7rij := Xij / x++ are Dirichlet distributed, if each Xij follows a Gamma distribution. See [PFTV92] how to sample from a Gamma distribution. The variance has been expanded in T~S, so the relative error Var [I]app" o.-Var[I] .. act of the approximation (11) and (18) are of the order of T'S and Var[Il e? act n (T~S)2 respectively, if z and J are dependent. If they are independent the leading term (11) drops itself down to order n -2 resulting in a reduced relative accuracy O( T~S) of (18). Comparison with the Monte Carlo values confirmed an accurracy in the range (T~S)1...2. The mean (4) is exact. Together with the skewness and kurtosis we have a good description for the distribution of the mutual information p(Iln) for not too small sample bin sizes nij' We want to conclude with some notes on useful accuracy. The hypothetical prior sample sizes n~j = {a, -!S' ~,1} can all be argued to be non-informative [GCSR95]. Since the central moments are expansions in n- 1 , the next to leading order term can be freely adjusted by adjusting n~j E [0 ... 1]. So one may argue that anything beyond leading order is free to will, and the leading order terms may be regarded as accurate as we can specify our prior knowledge. On the other hand, exact expressions have the advantage of being safe against cancellations. For instance, leading order of E [I] and E[I2] does not suffice to compute the leading order of Var[I]. Acknowledgements I want to thank Ivo Kwee for valuable discussions and Marco Zaffalon for encouraging me to investigate this topic. This work was supported by SNF grant 200061847.00 to Jiirgen Schmidhuber. References [AS74] M. Abramowitz and 1. A. Stegun, editors. Handbook of mathematical functions. Dover publications, inc., 1974. [Bra99] M. Brand. Structure learning in conditional probability models via an entropic prior and parameter extinction. N eural Computation, 11(5):1155- 1182, 1999. [Bun96] W. Buntine. A guide to the literature on learning probabilistic networks from data. IEEE Transactions on Knowledge and Data Engineering, 8:195- 210, 1996. [CT91] T. M. Cover and J. A. Thomas. Elements of Information Theory. Wiley Series in Telecommunications. John Wiley & Sons, New York, NY, USA, 1991. [GCSR95] A. Gelman, J. B. Carlin, H. S. Stern, and D. B. Rubin. Bayesian Data Analysis. Chapman, 1995. Learnig in [Hec98] D. Heckerman. A tutorial on learning with Bayesian networks. Graphical Models, pages 301-354, 1998. [KJ96] G. D. Kleiter and R. Jirousek. Learning Bayesian networks under the control of mutual information. Proceedings of the 6th International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems (IPMU-1996), pages 985- 990, 1996. [Kle99] G. D. Kleiter. The posterior probability of Bayes nets with strong dependences. Soft Computing, 3:162- 173, 1999. [PFTV92] W. H. Press, B. P. Flannery, S. A. Teukolsky, and W. T. Vetterling. Numerical R ecipes in C: Th e Art of Scientific Computing. Cambridge University Press, Cambridge, second edition, 1992. [SooOO] E. S. Soofi. Principal information theoretic approaches. Journal of the American Statistical Association, 95:1349- 1353, 2000. [WW93] D. R. Wolf and D. H. Wolpert. Estimating functions of distributions from A finite set of samples, part 2: Bayes estimators for mutual information, chisquared, covariance and other statistics. Technical Report LANL-LA-UR-93833, Los Alamos National Laboratory, 1993. Also Santa Fe Insitute report SFI-TR-93-07-047.
2071 |@word version:1 briefly:1 extinction:1 simulation:2 covariance:7 tr:1 moment:8 contains:1 series:1 ours:1 z2:1 john:1 numerical:4 informative:6 noninformative:1 drop:3 half:1 ivo:1 inspection:1 dover:1 short:1 node:1 mathematical:1 multinomially:1 incorrect:1 advocate:1 fitting:1 encouraging:1 estimating:1 bounded:1 suffice:1 what:1 skewness:8 hypothetical:1 act:2 exactly:1 wrong:1 scaled:1 control:1 grant:1 before:1 positive:1 engineering:1 limit:4 suggests:2 co:1 abramowitz:1 range:3 statistically:1 practical:1 procedure:1 logz:2 snf:1 empirical:1 suggest:1 get:11 cannot:1 gelman:1 context:1 www:1 dz:1 go:1 starting:1 independently:1 oste:1 estimator:1 utilizing:2 regarded:1 imax:3 exact:13 element:1 idsia:3 jirousek:1 rij:22 ensures:2 connected:1 jrc:1 valuable:1 ij2:1 depend:1 algebra:1 completely:1 manno:1 po:1 joint:3 represented:2 various:1 derivation:1 fast:1 insitute:1 monte:4 outcome:1 quite:1 heuristic:1 larger:1 widely:1 otherwise:1 cov:1 statistic:1 itself:1 ip:1 final:1 advantage:1 rr:2 net:4 kurtosis:8 analytical:1 rln:8 fr:5 inserting:3 j2:2 zaffalon:1 description:1 los:1 double:2 converges:2 derive:2 depending:1 ij:17 strong:1 dividing:1 involves:1 indicate:1 switzerland:1 concentrate:2 fij:7 safe:1 correct:1 stochastic:2 bin:1 argued:1 generalization:1 adjusted:1 strictly:2 correction:3 marco:1 around:4 normal:2 entropic:2 numb:1 a2:1 estimation:1 create:1 always:2 gaussian:4 modified:1 rather:1 ej:1 og:3 publication:1 derived:2 indicates:1 dependent:2 vetterling:1 expand:4 comprising:1 issue:2 logn:2 art:1 mutual:22 marginal:2 field:1 construct:1 haldane:3 equal:1 sampling:1 chapman:1 cancel:2 peaked:3 simplex:1 t2:1 others:1 report:2 gamma:7 national:1 n1:1 ab:2 interest:1 investigate:1 introduces:1 accurate:4 integral:3 unless:1 logarithm:2 nij:21 hutter:1 instance:6 soft:1 cover:1 deviation:3 rare:1 euler:1 uniform:3 comprised:1 alamo:1 too:1 buntine:1 dependency:2 answer:3 cxp:1 st:1 density:3 international:1 systematic:3 probabilistic:1 ansatz:1 together:1 quickly:2 again:1 central:8 management:1 containing:1 stochastically:1 american:1 derivative:2 leading:16 logr:1 account:1 lookup:1 inc:1 ad:1 later:1 bayes:2 complicated:1 il:1 ir:2 accuracy:3 ni:8 variance:26 bayesian:9 ijkl:1 carlo:4 confirmed:1 app:1 cumbersome:2 against:1 frequency:6 jiirgen:1 di:2 galleria:1 adjusting:1 knowledge:3 confluent:1 higher:2 specify:1 improved:1 done:1 strongly:3 furthermore:1 just:2 hand:2 scientific:1 usa:1 validity:2 fooo:1 true:2 hence:1 sfi:1 laboratory:1 i2:2 iln:15 anything:1 criterion:1 theoretic:2 fi:5 common:1 jl:1 discussed:2 he:1 association:1 significant:1 cambridge:2 rd:2 z4:1 similarly:1 cancellation:1 reliability:1 posterior:12 irrelevant:1 schmidhuber:1 arbitrarily:4 iiit:1 greater:1 additional:1 freely:1 determine:1 ii:1 technical:1 cross:1 variant:1 expectation:3 normalization:1 ion:2 whereas:1 want:2 bik:1 integer:4 noting:1 independence:1 carlin:1 gave:1 perfectly:1 computable:2 expression:19 speaking:1 york:1 jj:7 useful:1 santa:1 involve:1 reduced:1 http:1 xij:2 restricts:1 tutorial:1 judging:1 estimated:1 delta:1 correctly:1 per:1 discrete:1 shall:1 threshold:1 drawn:1 chisquared:1 asymptotically:1 sum:3 uncertainty:1 telecommunication:1 perk:3 informat:1 bound:1 quadratic:1 precisely:1 kronecker:1 ri:4 x2:1 argument:1 min:2 expanded:2 combination:1 kd:1 heckerman:1 son:1 sam:1 ur:1 wi:1 making:1 jeffreys:3 restricted:1 ln:5 discus:1 needed:2 know:1 available:2 away:1 appropriate:2 thomas:1 denotes:2 dirichlet:12 assumes:1 mam:1 graphical:1 log2:1 objective:1 question:3 already:1 quantity:2 dependence:2 thank:1 me:1 topic:1 argue:2 marcus:3 code:1 index:2 pie:1 unfortunately:1 fe:1 relate:1 negative:2 numerics:1 implementation:1 stern:1 unknown:3 upper:1 finite:3 t:1 sharp:1 arbitrary:1 pair:1 kl:1 lanl:1 unequal:1 hypergeometric:1 dts:1 beyond:2 usually:1 reliable:2 power:1 natural:1 lij:1 prior:28 literature:1 acknowledgement:1 relative:2 mixed:1 proportional:1 var:14 consistent:2 rubin:1 principle:2 editor:1 course:1 supported:1 last:1 soon:1 free:1 rior:1 allow:1 guide:1 taking:2 differentiating:1 distributed:2 calculated:1 stegun:1 commonly:2 clue:1 refinement:1 c5:3 transaction:1 approximate:2 keep:1 handbook:1 conclude:1 table:1 delving:1 rearranging:1 expanding:1 expansion:7 domain:1 noise:1 edition:1 eural:1 representative:1 quantiles:1 ny:1 wiley:2 ct91:2 comprises:2 lugano:1 bij:1 theorem:2 down:1 specific:1 er:1 entropy:1 flannery:1 lt:1 wolpert:1 expressed:2 ch:3 ipmu:1 wolf:1 chance:3 teukolsky:1 conditional:1 replace:1 fkl:2 determined:2 principal:1 called:1 total:1 la:1 brand:2 ex:2
1,176
2,072
Sampling Techniques for Kernel Methods Dimitris Achlioptas Microsoft Research optas@microsoft.com Frank McSherry University of Washington mcsherry@cs.washington.edu Bernhard Sch?olkopf Biowulf Technologies NY bs@conclu.de Abstract We propose randomized techniques for speeding up Kernel Principal Component Analysis on three levels: sampling and quantization of the Gram matrix in training, randomized rounding in evaluating the kernel expansions, and random projections in evaluating the kernel itself. In all three cases, we give sharp bounds on the accuracy of the obtained approximations. Rather intriguingly, all three techniques can be viewed as instantiations of the following idea: replace the kernel function by a ?randomized kernel? which behaves like in expectation. 1 Introduction Given a collection  of training data     , techniques such as linear SVMs [13] and PCA extract features from  by computing linear functions of this data. However, it is often the case that the structure present in the training data is not simply a linear function of the data representation. Worse, many data sets do not readily support linear operations such as addition and scalar multiplication (text, for example). In a ?kernel method?  is first mapped into some dot product space  using  . The dimension of  can be very large, even infinite, and therefore it may not be practical (or possible) to work with the mapped data explicitly. Nonetheless, in many cases the dot products  ! " can be evaluated efficiently using a positive definite kernel for  , ?.e. a function so that #$ %'&()*+! " . Any algorithm whose operations can be expressed in terms of dot products can be generalized to an algorithm which operates on , , simply by presenting the Gram matrix - . /&   $  as the input covariance matrix. Note that at no point is the function  explicitly computed; the kernel implicitly performs the dot product calculations between mapped points. While this ?kernel trick? has- been extremely successful, a problem common to all kernel is a dense matrix, making the input size scale as 021 . For methods is that, in general, example, in Kernel PCA such a matrix has to be diagonalized, while in SVMs a quadratic program of size 0 1 must be solved. As the size of training sets in practical applications increases, the growth of the input size rapidly poses severe computational limitations. Various methods have been proposed to deal with this issue, such as decomposition methods for SVM training (e.g., [10]), speedup methods for Kernel PCA [12], and other kernel methods [2, 14]. Our research is motivated by the need for such speedups that are also accompanied by strong, provable performance guarantees. In this paper we give three such speedups for Kernel PCA. We start by simplifying the Gram matrix via a novel matrix sampling/quantization scheme, motivated by spectral properties of random matrices. We then move on to speeding up classification, by using randomized rounding in evaluating kernel expansions. Finally, we consider the evaluation of kernel functions themselves and show how many popular kernels can be approximated efficiently. Our first technique relates matrix simplification to the stability of invariant subspaces. The other two are, in fact, completely general and apply to all kernel methods. What is more, our techniques suggest the notion of randomized kernels, whereby each evaluation of the kernel is replaced by an evaluation of a randomized function (on the same input pair). The idea is to use a function which for every input pair behaves like in expectation (over its internal coin-flips), yet confers significant computational benefits compared to using . In fact, each one of our three techniques can be readily cast as an appropriate randomized kernel, with no other intervention. 2 Kernel PCA Given 0 training points recall that is an 0  0 matrix with . &        . For some     method choice [11] computes the largest eigenvalues,     of  0 , the Kernel PCA (KPCA)  , and eigenvectors,        of . Then, given an input point  , the method  computes the value of nonlinear feature extraction functions   &    1        $#2 There are several methods for computing the principal components of a symmetric matrix. The choice depends on the properties of the matrix and on how many components one is seeking. In particular, if relatively few principal components are required, as is the case in KPCA, Orthogonal Iteration is a commonly used method.1  Orthogonal Iteration     matrix with orthonormal columns. 1. Let  be a random 0  2. While not converged, do (a)  (b)  Orthonormalize   3. Return  It is worth looking closely at the complexity of performing Orthogonal Iteration on a matrix the computational bottleneck. The  . Step 1 can be done in   0  steps, making  stepis 2overwhelmed orthonormalization step 2b takes time  0 1  and by the cost of comput ing  in step 2a which, generally, takes   0 1  . The number of iterations of the while loop is a somewhat complicated issue, but one can prove that the ?error? in  (with respect to the true principal components) decreases exponentially with the number of iterations. All in all, the running time of Orthogonal Iteration scales linearly with the cost of the matrix multiplication  . If  is sparse, ?.e., if roughly one  out of every  entries of  is non-zero, then the matrix multiplication  costs   0 1  . - As mentioned earlier, the matrix used in Kernel PCA is almost never sparse. In the next section, we will show how to sample and quantize the entries of , obtaining a matrix which is sparser and whose entries have simpler data representation, yet has essentially the same spectral structure, i.e. eigenvalues/eigenvectors, as . 1 Our discussion applies equally well to Lanczos Iteration which, while often preferable, is a more complicated method. Here we focus on Orthogonal Iteration to simplify exposition. 3 Sampling Gram Matrices In this section we describe two general ?matrix simplification? techniques and discuss their implications for Kernel PCA. In particular, under natural assumptions on the spectral structure of , we will prove that applying KPCA to the simplified matrix yields subspaces which are very close to those that KPCA would find in . As a result, when we project vectors onto these spaces (as performed by the feature extractors) the results are provably close to the original ones. First, our sparsification process works by randomly omitting entries in we let the matrix be described entrywise as  with probability - / & $  - &  .  - . Precisely stated,     , where     with probability    - .    2     with probability - Second, - our quantization process rounds each entry in to one of   . , thus reducing the representation of each entry to a single bit.  - .    with probability - / & $ & & Sparsification greatly accelerates the computation of eigenvectors by accelerating multiplication by . Moreover, both approaches greatly reduce the space required to store the matrix (and they can be readily combined), allowing for much bigger training sets to fit in main memory. Finally, we note that i) sampling also speeds up the construction of the that remain in , while ii) Gram matrix since we need only compute those values of quantization allows us to replace exact kernel evaluations by coarse unbiased estimators, which can be more efficient to compute.    While the two processes above are quite different, they share one important commonality: & in each case,  .  & , / . Moreover, the entries of the error matrix, are independent random variables, having expectation zero and bounded variance. Large deviation extensions [5] of Wigner?s famous semi-circle law, imply that with very high probability such matrices have small L2 norm (denoted by throughout).   !  Theorem 1 (Furedi and Komlos [5]) Let be an 0  0 symmetric matrix whose entries are independent random variables with mean 0, variance bounded above   , by 1 , and magnitude bounded by 0 0 . With probability  1 0 "$# &% '()     1  2"# 0   + * - , ./" 0 " " It is worth noting that this upper bound is within a constant factor of the lower bound on the L2 norm of any matrix where the mean squared entry equals 1 . More precisely, it is 0 . easy to show that every matrix with Frobenius norm  #0 $1 has L2 norm at least Therefore, we see that the L2 error introduced by is within a factor of 4 of the L2 error associated with any modification to that has the same entrywise mean squared error. 3" "4# We will analyze three different cases of spectral stability, corresponding to progressively stronger assumptions. At the heart of these results is the stability of invariant subspaces in the presence of additive noise. This stability is very strong, but can be rather technical to express. In stating each of these results, it is important to note that the eigenvectors correspond exactly to the feature extractors associated with Kernel PCA. For an input point  , let denote the vector whose th coordinate is  #$   and recall that 5 6  # &    1      5%87  .5  "2   - Recall that in KPCA we associate the largest eigenvalues of , where is  features   with  . First, we typically chosen  by requiring , for some threshold consider what     is not large. Observe that in this case we cannot hope happens when to equate    all # and   , as the th feature is very sensitive to small changes in . However,  far from  are treated consistently in - and - . we can show that all features with   Theorem 2 Let    be any matrix- whose columns form an orthonormal basis   for the space of features (eigenvectors) in whose eigenvalue is at least . Let    be any  matrix whose columns form an orthonormal basis for the orthogonal complement of    .    Let    and    be the analogous matrices for . For any ,             1      1 and     -              1   If we use the threshold for the eigenvalues of - , the first equation asserts that the features  . KPCA recovers are not among the features of whose eigenvalues are less than Similarly, the second equation that KPCA will recover all the features of whose  asserts eigenvalues are larger than .    Proof: We employ the techniques of Davis and Kahan [4]. Observe that -  $  & -   &           -         -    &                        &         and  are diagonal matrices whose entries (the eigenvalues of where  , respectively. Therefore least and at most                                 1      1  1           1  and  - ) are at       which implies the first stated result. The second proof is essentially identical. In our second result we will still not be able to isolate individual as the error and features,   . However, matrix can reorder their importance by, say, interchanging we can show that any such interchange will occur consistently in all test vectors. Let   be the  -dimensional vector whose  th coordinate is    1  # , ?.e., here we do not normalize features to ?equal importance?. Recall that is the vector whose th coordinate is     # . 5 6  1   1 for some   . There is an orthonormal    #& ! # 1   5            and  &  . Proof: Instantiate Theorem 2 with &     Theorem 3 Assume that  rotation matrix  such that for all  Note that the rotation matrix becomes completely irrelevant if we are only concerned with differences, angles, or inner products of feature vectors. Finally, we prove that in the special - case where a feature is well separated from its neighboring features in the spectrum of , we get a particularly strong bound. Theorem 4 If   ,       1   #+  1 , and           1      , then  #  5  - - Proof:(sketch) As before, we specialize Theorem 2, but first shift both and by in isolation. This does not change the eigenvectors, and allows us to consider    . 4 Approximating Feature Extractors Quickly Having determined eigenvalues and eigenvectors, given an input point  , the value of  on each feature reduces to evaluating, for some unit vector  , a function    &       $  2    ! 1 . Assume that   . where we dropped the subscript  , as by  well as the scaling take values in an interval of width and let  be any unit vector. We will devise a fast, unbiased, small-variance estimator for , by sampling and rounding the expansion coefficients   . Fix    1 1 6       (  then let   &   ; if        let    with probability !     . For each : if  &   otherwise. That is, after potentially keeping some large coefficients deterministically, we proceed to perform ?randomized rounding? on the (remaining) coefficients of  . Let  Clearly, we have    & bound the behavior of  # ing. In particular, this gives # &         $*   # . Moreover, using Hoeffding?s inequality [7], we can    #  arising from the terms subjected to probabilistic round    +1   +* -,    #&     0  1  (1)  Note now that in Kernel PCA we typically expect    # 0 , i.e., dense eigenvectors.  This makes # 0 the natural scale for measuring  # and suggests that using far fewer than 0 kernel evaluations we can get good approximations of chosen (fixed) value of  let us say that  # is trivial if       % '( 0   . In particular, for a  Having picked some threshold  (for SVM expansions  is related to the classification offset) we want to determine whether   is non-trivial and, if so, we want to get a good relative error estimate for it.    and  2   0 1. There are fewer than      ) Theorem 5 For any  probability at least 2. Either both  0  % #      0 0 set  &  # 0    . With non-zero   .    are trivial or       %  and +%1'(             . Note that 0  Proof:  Let 0 denote the number of non-zero   and let &  6       equals   plus the sum of 0   independent Bernoulli trials. It is not hard to show that the probability that the event  in 1 fails is bounded by the corresponding probability for the case where all coordinates of  are equal.  In that case, 0 is a Binomial random variable   0   with 0 trials and probability of success  # 0 and, by our choice of  ,   0  &   )  . The Chernoff bound now implies that the event in 1 fails to occur with probability ! 0       is at least . For the enent in 2 it suffices to observe that failure occurs only if    % '( 0 . By (1), this also occurs with probability ! 0 )  .   " #  5 Quick batch approximations of Kernels In this section we devise fast approximations of the kernel function itself. We focus on kernels sharing the following two characteristics: i) they map -dimensional Euclidean space, and, ii) the mapping depends only on the distance and/or inner product of the considered points. We note that this covers some of the most popular kernels, e.g., RBFs and polynomial kernels. To simplify exposition we focus on the following task: given a sequence of (test) vectors        determine      for each of a fixed set of (training) vectors 1      , where 0 . To get a fast batch approximdition, the idea is that rather than evaluating distances and inner products directly, we will use a fast, approximately correct oracle for these quantities offering the following guarantee: it will answer all queries with small relative error. A natural approach for creating such an oracle is to pick  of the coordinates in input space and use the projection onto these coordinates to determine distances  and  inner products. The problem with this approach is that if    &              , any coordinate sampling scheme is bound to do poorly. On the other hand, if we knew that all coordinates contributed ?approximately equally? to      , then coordinate sampling would be much more appealing. We will do just this, using the technique of random projections [8], which can be viewed as coordinate sampling preceded by a random rotation.       Imagine that we applied a spherically random rotation  to !    (before training) and then applied the same random rotation  to each input point   as it became available. Clearly, all distances and inner products would remain the same and we would get exactly the same results as without the rotation. The interesting part is that any fixed vector that was a linear combination of training and/or input vectors, e.g.   , after being rotated  becomes a spherically random vector of length . As a result, the coordinates of are    ! , enabling coordinate sampling. i.i.d. random variables, in fact   $ #  Our oracle amounts to multiplying %1'( each training and input point by the same   projection matrix , where  &    , and using the resulting  -dimensional points to estimate rotation matrix distances and inner products. (Think of as the result of taking a  and keeping the first  columns (sampling)). Before describing the choice of and the quality of the resulting approximations, let us go over the computational savings. %1'(  . Note that 1. Rotating the 0 training vectors takes   0  This cost will be amortized over the sequence of input vectors.  This rotation can be performed in the training phase. 2. The kernel evaluations for each   now take  0  instead of  0 . 3. Rotating  takes time    which is dominated by  0  . %1'( % '( %1'( Having motivated our oracle as a spherically random rotation followed by coordinate sampling, we will actually employ a simpler method to perform the projection. Namely, we will rely on a recent result of [1], asserting that we can do at least as well by taking  * where the */ are i.i.d. with  /  , each case having probabil / & .    until the end, ity . Thus, postponing the scaling by each of the  new coordinates is formed as follows: split the coordinates randomly into two groups; sum the coordinates in each group; take the difference of the two sums.  #        # Regarding the quality of approximations we get Theorem 6 Consider sets of points   any and for given   let  &    and #%    in  . Let  & 0 2  %   1   )      Let .  be a random   matrix defined by .  , each case having probability       the   &  .*For/ # any where    let pair of points #        , for every   1  .       1  /           1      "&    "     1   With probability at least  and     1  (2)  (3)  ,  +  *      ,    )            Proof: We use Lemma 5 of [1], asserting that for any      1       1  . are i.i.d. with denote .  and any   1       1       (4)     1   . Thus, by the By our choice of  , the r.h.s. of (4) is union bound, with prob   the lengths of all 0  0   vectors ability at least corresponding to  1 a factor of  . This    and    , &    ,  &  0 , are maintained within    1      1 and thus if readily we observe that      " &    1    1    yields  1 and(2).  For (3)  1 are   within  of  1   1 and   1 , then (3) holds.     6   6 Conclusion We have described three techniques for speeding up kernel methods through the use of randomization. While the discussion has focused on Kernel PCA, we feel that our techniques have potential for further development and empirical evaluation in a more general setting. Indeed, the methods for sampling kernel expansions and for speeding up the kernel evaluation are universal; also, the Gram matrix sampling is readily applicable to any kernel technique based on the eigendecomposition of the Gram matrix [3]. Furthermore, it might enable us to speed up SVM training by sparsifying the Hessian and then applying a sparse QP solver, such as the ones described in [6, 9]. Our sampling and quantization techniques, both in training and classification, amount to repeatedly replacing single kernel evaluations with independent random variables that have appropriate expectations. Note, for example, that while we have represented the sampling of the kernel expansion as randomized rounding of coefficients, this rounding is also equivalent to the following process: consider each coefficients as is, but replace every kernel invocation   ! with an invocation of a randomized kernel function, distributed as   $  $ &       with probability  otherwise. Similarly, the process of sampling in training can be thought of as replacing      &        with probability with probability    with  while an analogous randomized kernel is the obvious choice for quantization. We feel that this approach suggests a notion of randomized kernels, wherein kernel evaluations are no longer considered as deterministic but inherently random, providing unbiased estimators for the corresponding inner products. Given bounds on the variance of these estimators, it seems that algorithms which reduce to computing weighted sums of kernel evaluations can exploit concentration of measure. Thus, randomized kernels appear promising as a general tool for speeding up kernel methods, warranting further investigation. Acknowledgments. BS would like to thank Santosh Venkatesh for detailed discussions on sampling kernel expansions. References [1] D. Achlioptas, Database-friendly random projections, Proc. of the 20th Symposium on Principle of Database Systems (Santa Barbara, California), 2001, pp. 274?281. [2] C. J. C. Burges, Simplified support vector decision rules, Proc. of the 13th International Conference on Machine Learning, Morgan Kaufmann, 1996, pp. 71?77. [3] N. Cristianini, J. Shawe-Taylor, and H. Lodhi, Latent semantic kernels, Proc. of the 18th International Conference on Machine Learning, Morgan Kaufman, 2001. [4] C. Davis and W. Kahan, The rotation of eigenvectors by a perturbation 3, SIAM Journal on Numerical Analysis 7 (1970), 1?46. [5] Z. F?uredi and J. Koml?os, The eigenvalues of random symmetric matrices, Combinatorica 1 (1981), no. 3, 233?241. [6] N. I. M. Gould, An algorithm for large-scale quadratic programming, IMA Journal of Numerical Analysis 11 (1991), no. 3, 299?324. [7] W. Hoeffding, Probability inequalities for sums of bounded random variables, Journal of the American Statistical Association 58 (1963), 13?30. [8] W. B. Johnson and J. Lindenstrauss, Extensions of Lipschitz mappings into a Hilbert space, Conference in modern analysis and probability (New Haven, Conn., 1982), American Mathematical Society, 1984, pp. 189?206. [9] R. H. Nickel and J. W. Tolle, A sparse sequential quadratic programming algorithm, Journal of Optimization Theory and Applications 60 (1989), no. 3, 453?473. [10] E. Osuna, R. Freund, and F. Girosi, An improved training algorithm for support vector machines, Neural Networks for Signal Processing VII, 1997, pp. 276?285. [11] B. Sch?olkopf, A. J. Smola, and K.-R. M?uller, Nonlinear component analysis as a kernel eigenvalue problem, Neural Computation 10 (1998), 1299?1319. [12] A. J. Smola and B. Sch?olkopf, Sparse greedy matrix approximation for machine learning, Proc. of the 17th International Conference on Machine Learning, Morgan Kaufman, 2000, pp. 911?918. [13] V. Vapnik, The nature of statistical learning theory, Springer, NY, 1995. [14] C. K. I. Williams and M. Seeger, Using the Nystrom method to speed up kernel machines, Advances in Neural Information Processing Systems 2000, MIT Press, 2001.
2072 |@word trial:2 polynomial:1 seems:1 norm:4 stronger:1 lodhi:1 covariance:1 decomposition:1 simplifying:1 pick:1 offering:1 diagonalized:1 com:1 yet:2 must:1 readily:5 numerical:2 additive:1 girosi:1 progressively:1 greedy:1 instantiate:1 fewer:2 tolle:1 coarse:1 simpler:2 mathematical:1 symposium:1 prove:3 specialize:1 indeed:1 behavior:1 themselves:1 roughly:1 solver:1 becomes:2 project:1 moreover:3 bounded:5 what:2 kaufman:2 sparsification:2 guarantee:2 every:5 friendly:1 growth:1 preferable:1 exactly:2 unit:2 intervention:1 appear:1 positive:1 before:3 dropped:1 subscript:1 approximately:2 might:1 plus:1 warranting:1 suggests:2 practical:2 acknowledgment:1 union:1 definite:1 orthonormalization:1 universal:1 empirical:1 thought:1 projection:6 suggest:1 get:6 onto:2 close:2 cannot:1 applying:2 confers:1 map:1 quick:1 equivalent:1 deterministic:1 go:1 williams:1 focused:1 estimator:4 rule:1 orthonormal:4 ity:1 stability:4 notion:2 coordinate:16 analogous:2 feel:2 construction:1 imagine:1 exact:1 programming:2 trick:1 associate:1 amortized:1 approximated:1 particularly:1 database:2 solved:1 decrease:1 mentioned:1 complexity:1 cristianini:1 completely:2 basis:2 various:1 represented:1 separated:1 fast:4 describe:1 query:1 whose:12 quite:1 larger:1 say:2 otherwise:2 ability:1 kahan:2 think:1 itself:2 komlos:1 sequence:2 eigenvalue:11 propose:1 product:11 neighboring:1 loop:1 rapidly:1 poorly:1 frobenius:1 asserts:2 normalize:1 olkopf:3 probabil:1 rotated:1 stating:1 pose:1 uredi:1 strong:3 c:1 implies:2 closely:1 correct:1 enable:1 fix:1 suffices:1 investigation:1 randomization:1 extension:2 hold:1 considered:2 mapping:2 commonality:1 proc:4 applicable:1 sensitive:1 largest:2 tool:1 weighted:1 hope:1 uller:1 mit:1 clearly:2 rather:3 focus:3 consistently:2 bernoulli:1 greatly:2 seeger:1 typically:2 provably:1 issue:2 classification:3 among:1 denoted:1 development:1 special:1 equal:4 santosh:1 never:1 intriguingly:1 washington:2 sampling:18 extraction:1 having:6 identical:1 chernoff:1 saving:1 interchanging:1 simplify:2 haven:1 few:1 employ:2 modern:1 randomly:2 individual:1 ima:1 replaced:1 phase:1 microsoft:2 evaluation:11 severe:1 mcsherry:2 implication:1 orthogonal:6 euclidean:1 taylor:1 rotating:2 circle:1 column:4 earlier:1 cover:1 lanczos:1 measuring:1 kpca:7 cost:4 deviation:1 entry:10 successful:1 rounding:6 johnson:1 answer:1 combined:1 international:3 randomized:13 siam:1 probabilistic:1 quickly:1 squared:2 hoeffding:2 worse:1 creating:1 american:2 return:1 potential:1 de:1 accompanied:1 coefficient:5 explicitly:2 depends:2 performed:2 picked:1 analyze:1 start:1 recover:1 complicated:2 rbfs:1 formed:1 accuracy:1 became:1 variance:4 characteristic:1 efficiently:2 equate:1 yield:2 correspond:1 kaufmann:1 famous:1 multiplying:1 worth:2 converged:1 sharing:1 failure:1 nonetheless:1 pp:5 obvious:1 nystrom:1 associated:2 proof:6 recovers:1 popular:2 recall:4 hilbert:1 actually:1 wherein:1 improved:1 entrywise:2 done:1 evaluated:1 furthermore:1 just:1 smola:2 achlioptas:2 until:1 sketch:1 hand:1 replacing:2 nonlinear:2 o:1 quality:2 omitting:1 requiring:1 true:1 unbiased:3 symmetric:3 spherically:3 semantic:1 deal:1 round:2 width:1 davis:2 whereby:1 maintained:1 generalized:1 presenting:1 performs:1 wigner:1 novel:1 common:1 rotation:10 behaves:2 preceded:1 qp:1 exponentially:1 association:1 significant:1 similarly:2 shawe:1 dot:4 longer:1 recent:1 irrelevant:1 barbara:1 store:1 inequality:2 success:1 devise:2 morgan:3 somewhat:1 determine:3 signal:1 semi:1 relates:1 ii:2 reduces:1 ing:2 technical:1 calculation:1 equally:2 bigger:1 essentially:2 expectation:4 iteration:8 kernel:55 addition:1 want:2 interval:1 sch:3 isolate:1 noting:1 presence:1 split:1 easy:1 concerned:1 fit:1 isolation:1 reduce:2 idea:3 inner:7 regarding:1 shift:1 bottleneck:1 whether:1 motivated:3 pca:11 accelerating:1 proceed:1 hessian:1 repeatedly:1 generally:1 detailed:1 eigenvectors:9 santa:1 amount:2 svms:2 arising:1 express:1 group:2 sparsifying:1 conn:1 threshold:3 sum:5 angle:1 prob:1 almost:1 throughout:1 decision:1 scaling:2 bit:1 accelerates:1 bound:9 followed:1 simplification:2 quadratic:3 oracle:4 occur:2 precisely:2 dominated:1 speed:3 extremely:1 performing:1 relatively:1 gould:1 speedup:3 combination:1 remain:2 osuna:1 appealing:1 b:2 making:2 modification:1 happens:1 invariant:2 heart:1 equation:2 discus:1 describing:1 flip:1 subjected:1 end:1 koml:1 available:1 operation:2 apply:1 observe:4 appropriate:2 spectral:4 batch:2 coin:1 original:1 binomial:1 running:1 remaining:1 exploit:1 approximating:1 society:1 seeking:1 move:1 quantity:1 occurs:2 concentration:1 biowulf:1 diagonal:1 subspace:3 distance:5 thank:1 mapped:3 trivial:3 provable:1 length:2 providing:1 postponing:1 potentially:1 frank:1 stated:2 perform:2 allowing:1 upper:1 contributed:1 enabling:1 looking:1 orthonormalize:1 perturbation:1 sharp:1 introduced:1 complement:1 pair:3 cast:1 required:2 namely:1 venkatesh:1 california:1 able:1 dimitris:1 program:1 memory:1 event:2 natural:3 treated:1 rely:1 scheme:2 technology:1 imply:1 extract:1 speeding:5 text:1 l2:5 multiplication:4 relative:2 law:1 freund:1 expect:1 interesting:1 limitation:1 nickel:1 eigendecomposition:1 principle:1 share:1 keeping:2 burges:1 taking:2 sparse:5 benefit:1 distributed:1 dimension:1 gram:7 evaluating:5 lindenstrauss:1 computes:2 asserting:2 interchange:1 collection:1 commonly:1 simplified:2 far:2 implicitly:1 bernhard:1 instantiation:1 reorder:1 knew:1 spectrum:1 latent:1 promising:1 nature:1 inherently:1 obtaining:1 expansion:7 quantize:1 dense:2 main:1 linearly:1 noise:1 ny:2 fails:2 deterministically:1 comput:1 invocation:2 extractor:3 theorem:8 offset:1 svm:3 quantization:6 vapnik:1 sequential:1 importance:2 magnitude:1 overwhelmed:1 sparser:1 vii:1 simply:2 expressed:1 scalar:1 applies:1 springer:1 viewed:2 exposition:2 replace:3 lipschitz:1 change:2 hard:1 infinite:1 determined:1 operates:1 reducing:1 principal:4 lemma:1 internal:1 support:3 combinatorica:1
1,177
2,073
A Natural Policy Gradient Sham Kakade Gatsby Computational Neuroscience Unit 17 Queen Square, London, UK WC1N 3AR http: //www.gatsby.ucl.ac.uk sham @gatsby.ucl.ac.uk Abstract We provide a natural gradient method that represents the steepest descent direction based on the underlying structure of the parameter space. Although gradient methods cannot make large changes in the values of the parameters, we show that the natural gradient is moving toward choosing a greedy optimal action rather than just a better action. These greedy optimal actions are those that would be chosen under one improvement step of policy iteration with approximate, compatible value functions, as defined by Sutton et al. [9]. We then show drastic performance improvements in simple MDPs and in the more challenging MDP of Tetris. 1 Introduction There has been a growing interest in direct policy-gradient methods for approximate planning in large Markov decision problems (MDPs). Such methods seek to find a good policy 7r among some restricted class of policies, by following the gradient of the future reward. Unfortunately, the standard gradient descent rule is noncovariant. Crudely speaking, the rule !:l.()i = oJ] f / a()i is dimensionally inconsistent since the left hand side has units of ()i and the right hand side has units of l/()i (and all ()i do not necessarily have the same dimensions). In this paper, we present a covariant gradient by defining a metric based on the underlying structure of the policy. We make the connection to policy iteration by showing that the natural gradient is moving toward choosing a greedy optimal action. We then analyze the performance of the natural gradient in both simple and complicated MDPs. Consistent with Amari's findings [1], our work suggests that the plateau phenomenon might not be as severe using this method. 2 A Natural Gradient A finite MDP is a tuple (S, So, A, R, P) where: S is finite set of states, So is a start state, A is a finite set of actions, R is a reward function R : S x A --+ [0, Rmax], and P is the transition model. The agent 's decision making procedure is characterized by a stochastic policy 7r(a; s) , which is the probability of taking action a in state s (a semi-colon is used to distinguish the random variables from the parameters of the distribution). We make the assumption that every policy 7r is ergodic, ie has a well-defined stationary distribution p7f. Under this assumption, the average reward (or undiscounted reward) is 1]( 7r) == 2:: s ,a p7f (s )7r(a; S)R(s, a), the state-action value is Q7f(S, a) == E7f{2:::oR(st,at) -1](7r)lso = s,ao = a} and the value function is J7f(s) == E7f(a' ;s) {Q7f(s, a')}, where and St and at are the state and action at time t. We consider the more difficult case where the goal of the agent is to find a policy that maximizes the average reward over some restricted class of smoothly parameterized policies, fr = {7rO : 8 E ~m}, where tro represents the policy 7r(a; S, 8). The exact gradient of the average reward (see [8, 9]) is: \11](8) = Lp7f(s)\17r(a;s, 8)Q7f(s ,a) (1) s,a where we abuse notation by using 1](8) instead of 1](7ro). The steepest descent direction of 1](8) is defined as the vector d8 that minimizes 1](8 + d8) under the constraint that the squared length Id812 is held to a small constant. This squared length is defined with respect to some positive-definite matrix G(8), ie Id812 == 2::ij Gij (8)d8 i d8j = d8 T G(8)d8 (using vector notation). The steepest descent direction is then given by G- 1 \11](8) [1]. Standard gradient descent follows the direction \11](8) which is the steepest descent under the assumption that G(8) is the identity matrix, I. However, this as hoc choice of a metric is not necessarily appropriate. As suggested by Amari [1], it is better to define a metric based not on the choice of coordinates but rather on the manifold (ie the surface) that these coordinates parameterize. This metric defines the natural gradient. Though we slightly abuse notation by writing 1](8), the average reward is technically a function on the set of distributions {7rO : 8 E ~m}. To each state s, there corresponds a probability manifold, where the distribution 7r(a; S, 8) is a point on this manifold with coordinates 8. The Fisher information matrix of this distribution 7r(a; s,8) is F (8) = E s - 7f(a;s,O) [81o g 7r(a; s,8) 08 i olog 7r(a; s,8)] 08 ' j (2) and it is clearly positive definite. As shown by Amari (see [1]), the Fisher information matrix, up to a scale, is an invariant metric on the space of the parameters of probability distributions. It is invariant in the sense that it defines the same distance between two points regardless of the choice of coordinates (ie the parameterization) used, unlike G = I. Since the average reward is defined on a set of these distributions , the straightforward choice we make for the metric is: (3) where the expectation is with respect to the stationary distribution of 7ro. Notice that although each Fs is independent of the parameters of the MDP's transition model, the weighting by the stationary distribution introduces dependence on these parameters. Intuitively, Fs (8) measures distance on a probability manifold corresponding to state sand F(8) is the average such distance. The steepest descent direction this gives is: (4) 3 The Natural Gradient and Policy Iteration We now compare policy improvement under the natural gradient to policy iteration. For an appropriate comparison, consider the case in which Q7r (s, a) is approximated by some compatible function approximator r(s ,a;w) parameterized by w [9, 6]. 3.1 Compatible Function Approximation For vectors (), w E ~m, we define: 'IjJ (s , a)7r = \7logn(a;s,()), r(s,a;w) = wT 'ljJ7r(s,a) (5) where [\7logn(a ;s, ())]i = 8logn(a;s, ())!8()i. Let w minimize the squared error f(W, n) == L,s ,a p7r (s )n(a; s, ())(r (s, a; w) _Q7r (s, a))2. This function approximator is compatible with the policy in the sense that if we use the approximations f7r (s, a; w) in lieu of their true values to compute the gradient (equation 1), then the result would still be exact [9, 6] (and is thus a sensible choice to use in actor-critic schemes). Theorem 1. Let w minimize the squared error f(W, no). Then w= ~1}(()) . Proof. Since w minimizes the squared error, it satisfies the condition 8f!8wi = 0, which implies: LP7r(s)n(a;s,())'ljJ7r (s,a)('ljJ7r (s,a?w - Q7r(s,a)) = O. s,a or equivalently: s,a s,a By definition of 'ljJ7r, \7n(a;s,()) = n(a;s,())'ljJ7r(s,a) and so the right hand side is equal to \71}. Also by definition of 'ljJ7r, F( ()) = L,s ,a p7r (s )n( a; s, ()) 'ljJ7r (s, a)'ljJ7r (s, a) T. Substitution leads to: F(())w = \71}(()) . Solving for w gives w = F(()) - l\71}(()), and the result follows from the definition of the natural gradient. D Thus, sensible actor-critic frameworks (those using f7r(s , a; w)) are forced to use the natural gradient as the weights of a linear function approximator. If the function approximation is accurate, then good actions (ie those with large state-action values) have feature vectors that have a large inner product with the natural gradient. 3.2 Greedy Policy Improvement A greedy policy improvement step using our function approximator would choose action a in state s if a E argmaxa, f7r (s, a'; w). In this section, we show that the natural gradient tends to move toward this best action, rather than just a good action. Let us first consider policies in the exponential family (n(a ;s, ()) IX exp(()T?sa) where ?sa is some feature vector in ~m). The motivation for the exponential family is because it has affine geometry (ie the flat geometry of a plane), so a translation of a point by a tangent vector will keep the point on the manifold. In general, crudely speaking, the probability manifold of 7r(a; s, 0) could be curved, so a translation of a point by a tangent vector would not necessarily keep the point on the manifold (such as on a sphere). We consider the general (non-exponential) case later. We now show, for the exponential family, that a sufficiently large step in the natural gradient direction will lead to a policy that is equivalent to a policy found after a greedy policy improvement step. Theorem 2. For 7r(a; s, 0) ex: exp(OT 1>sa), assume that ~'TJ(O) is non-zero and that w minimizes the approximation error. Let7roo (a;s) =lima-+oo7r(a;s , O+a~'TJ(O)). Then 7r 00 (a; s) 1- 0 if and only if a E argmaxa, F' (s, a'; w). Proof. By the previous result, F'(s,a ;w) = ~'TJ(O)T'lj;7r(s,a). By definition of 7r(a; s, 0) , 'lj;7r (s, a) = 1>sa - E 7r (a';s ,O) (1)sa'). Since E 7r (a';s,O) (1)sa') is not a function of a, it follows that argmax a , r(s, a'; w) = argmax a , ~'TJ(Of 1>sa' . After a gradient step, 7r(a; s, 0 + a~'TJ(O)) ex: exp(OT 1>sa + a~'TJ(O)T 1>sa). Since ~'TJ(O) 1- 0, it is clear that as a -+ 00 the term ~'TJ(O)T 1>sa dominates , and so D 7r 00 (a, s) = 0 if and only if a f{. argmax a , ~ 'TJ( 0) T 1>sa' . It is in this sense that the natural gradient tends to move toward choosing the best action. It is straightforward to show that if the standard non-covariant gradient rule is used instead then 7r oo (a; s) will select only a better action (not necessarily the best), ie it will choose an action a such that F'(s ,a;w) > E 7r (a';s){F'(s,a';w)}. Our use of the exponential family was only to demonstrate this point in the extreme case of an infinite learning rate. Let us return to case of a general parameterized policy. The following theorem shows that the natural gradient is locally moving toward the best action, determined by the local linear approximator for Q7r (s, a). Theorem 3. Assume that w minimizes the approximation error and let the update to the parameter be 0' = 0 + a~'TJ(O). Then 7r(a; s, 0') = 7r(a; s, 0)(1 + r(s , a; w)) + 0(a 2 ) Proof. The change in 0, ,6.0, is a~'TJ(O), so by theorem 1, ,6.0 = aw. To first order, 7r(a; s, 0') 7r(a; s, 0) + fJ7r(a~;, O)T ,6.0 + 0(,6.0 2 ) 7r(a; s, 0)(1 7r(a; s, 0)(1 7r(a;s,O)(l + 'lj;(s, af ,6.0) + 0(,6.0 2 ) + a'lj;(s, af w) + 0(a 2 ) + ar(s,a;w)) + 0(a 2 ) , where we have used the definition of 'lj; and f. D It is interesting to note that choosing the greedy action will not in general improve the policy, and many detailed studies have gone into understanding this failure [3]. However, with the overhead of a line search, we can guarantee improvement and move toward this greedy one step improvement. Initial improvement is guaranteed since F is positive definite. 4 Metrics and Curvatures Obviously, our choice of F is not unique and the question arises as to whether or not there is a better metric to use than F. In the different setting of parameter estimation, the Fisher information converges to the Hessian, so it is asymptotically efficient [1], ie attains the Cramer-Rao bound. Our situation is more similar to the blind source separation case where a metric is chosen based on the underlying parameter space [1] (of non-singular matrices) and is not necessarily asymptotically efficient (ie does not attain second order convergence). As argued by Mackay [7], one strategy is to pull a metric out of the data-independent terms of the Hessian (if possible), and in fact, Mackay [7] arrives at the same result as Amari for the blind source separation case. Although the previous sections argued that our choice is appropriate, we would like to understand how F relates to the Hessian V 2 TJ(B), which, as shown in [5], has the form: sa (6) Unfortunately, all terms in this Hessian are data-dependent (ie are coupled to stateaction values) . It is clear that F does not capture any information from these last two terms, due to their VQ7r dependence. The first term might have some relation to F due the factor of V 2 7f. However, the Q values weight this curvature of our policy and our metric is neglecting such weighting. Similar to the blind source separation case, our metric clearly does not necessarily converge to the Hessian and so it is not necessarily asymptotically efficient (ie does not attain a second order convergence rate). However, in general, the Hessian will not be positive definite and so the curvature it provides could be of little use until B is close to a local maxima. Conjugate methods would be expected to be more efficient near a local maximum. 5 Experiments We first look at the performance of the natural gradient in a few simple MDPs before examining its performance in the more challenging MDP of Tetris. It is straightforward to estimate F in an online manner, since the derivatives V log 7f must be computed anyway to estimate VTJ(B). If the update rule f f- f + V log 7f(at; St,B)Vlog7f(at; St,Bf is used in a T-Iength trajectory, then fiT is a consistent estimate of F. In our first two examples, we do not concern ourselves with sampling issues and instead numerically integrate the exact derivative (B t = Bo + J~ VTJ(BddB). In all of our simulations, the policies tend to become deterministic (V log 7f -+ 0) and to prevent F from becoming singular, we add about 10- 3 1 at every step in all our simulations. We simulated the natural policy gradient in a simple I-dimensional linear quadratic regulator with dynamics x(t + 1) = .7x(t) + u(t) + E(t) and noise distribution E ~ G(O,l). The goal is to apply a control signal u to keep the system at x = 0, (incurring a cost of X(t)2 at each step). The parameterized policy used was 7f(u; x, B) ex exp(Blx 2 + B2X). Figure lA shows the performance improvement when the units of the parameters are scaled by a factor of 10 (see figure text). Notice that the time to obtain a score of about 22 is about three orders of magnitude - '--''''' '~''? ~...... unsealed - $=10 s=1 ...... 1 2 -'::',$,=1 $2=10 _. - ":. \: .:-. W ~2:=':3=::l4' --"-,-, L _-=-2 --': -'':::::;0:::=:::' I09 10 (time) (' ,\2 ~R=O) ~ rl I "E 1 Ir h "::>:" 20 B 8' ~a C 21 i D -11 ______ ',05 .. ~ ~0 '::0 --0~5C------:-'-----:-': '5C------::' 2 ~21 ~, 0 7 time x 10 L------------ a 0.5 1 /:--------1. - 1.5 time 2 2.5 3 " Q- "\" .\'; L---::-,::::;?7J~========-~ 5 8., 10 15 Figure 1: A) The cost Vs. 10glo(time) for an LQG (with 20 time step trajectories). The policy used was 7f(u; x, ()) ex: exp(()lslX2 + ()2S2X) where the rescaling constants, Sl and S2, are shown in the legend. Under equivalent starting distributions (()lSl = ()2S2 = -.8) , the right-most three curves are generated using the standard gradient method and the rest use the natural gradient. B) See text. C top) The average reward vs. time (on a 107 scale) of a policy under standard gradient descent using the sigmoidal policy parameterization (7f(I; s, ()i) ex: exp(()i)/(1 + exp(()i)), with the initial conditions 7f(i , 1) = .8 and 7f(j, 1) = .1. C bottom) The average reward vs. time (unscaled) under standard gradient descent (solid line) and natural gradient descent (dashed line) for an early window of the above plot. D) Phase space plot for the standard gradient case (the solid line) and the natural gradient case (dashed line) . faster. Also notice that the curves under different rescaling are not identical. This is because F is not an invariant metric due to the weighting by Ps. The effects of the weighting by p(s) are particularly clear in a simple 2-state MDP (Figure IB), which has self- and cross-transition actions and rewards as shown. Increasing the chance of a self-loop at i decreases the stationary probability of j. Using a sigmoidal policy parameterization (see figure text) and initial conditions corresponding to p(i) = .8 and p(j) = .2, both self-loop action probabilities will initially be increased under a gradient rule (since one step policy improvement chooses the self-loop for each state). Since the standard gradient weights the learning to each parameter by p(s) (see equation 1), the self-loop action at state i is increased faster than the self loop probability at j, which has the effect of decreasing the effective learning-rate to state j even further. This leads to an extremely fiat plateau with average reward 1 (shown in Figure lC top), where the learning for state j is thwarted by its low stationary probability. This problem is so severe that before the optimal policy is reached p(j) drops as low as 10- 7 from its initial value of .2, which is disastrous for sampling methods. Figure 1 C bottom shows the performance of the natural gradient (in a very early time window of Figure lC top). Not only is the time to the optimal policy decreased by a factor of 107 , the stationary distribution of state i never drops below .05. Note though the standard gradient does increase the average reward faster at the start, but only to be seduced by sticking at state i. The phase space plot in Figure ID shows the uneven learning to the different parameters, which is at the heart of the problem. In general, if a table lookup Boltzmann policy is used (ie 7f( a; s , ()) ex: exp( () sa)), it is straightforward to show that the natural gradient weights the components of ~'fJ uniformly (instead of using p(s)), thus evening evening out the learning to all parameters. The game of Tetris provides a challenging high dimensional problem. As shown in [3], greedy policy iteration methods using a linear function approximator exhibit drastic performance degradation after providing impressive improvement (see [3] for a description of the game, methods , and results). The upper curve in Figure2A replicates these results. Tetris provides an interesting case to test gradient methods, A 5000, - - - - - - - - - - - - - , B 7000, - - - - - - - - - - , - - - - , C 6000 4000 5000 ~3000 ~4000 ?0 a... 2000 &3000 2000 1000 1000 1 I09,O( lteralions) 2 500 1000 Iterations 1500 2000 Figure 2: A) Points vs. 10g(Iterations) . The top curve duplicates the same results in [3] using the same features (which were simple functions of the heights of each column and the number of holes in the game). We have no explanation for this performance degradation (nor does [3]). The lower curve shows the poor performance of the standard gradient rule. B) The curve on the right shows the natural policy gradient method (and uses the biased gradient method of [2] though this method alone gave poor performance). We found we could obtain faster improvement and higher asymptotes if the robustifying factor of 10- 3 I that we added to F was more carefully controlled (we did not carefully control the parameters). C) Due to the intensive computational power required of these simulations we ran the gradient in a smaller Tetris game (height of 10 rather than 20) to demonstrate that the standard gradient updates (right curve) would eventually reach the same performance of the natural gradient (left curve). which are guaranteed not to degrade the policy. We consider a policy compatible with the linear function approximator used in [3] (ie 7f(a ;s, (}) ex: exp((}T?sa) where ?sa are the same feature vectors). The features used in [3] are the heights of each column, the differences in height between adjacent columns, the maximum height, and the number of 'holes' . The lower curve in Figure 2A shows the particularly poor performance of the standard gradient method. In an attempt to speed learning, we tried a variety of more sophisticated methods to no avail, such as conjugate methods, weight decay, annealing, the variance reduction method of [2], the Hessian in equation 6, etc. Figure 2B shows a drastic improvement using the natural gradient (note that the timescale is linear). This performance is consistent with our theoretical results in section 3, which showed that the natural gradient is moving toward the solution of a greedy policy improvement step. The performance is somewhat slower than the greedy policy iteration (left curve in Figure 2B) which is to be expected using smaller steps. However, the policy does not degrade with a gradient method. Figure 2 shows that the performance of the standard gradient rule (right curve) eventually reaches the the same performance of the natural gradient, in a scaled down version of the game (see figure text). 6 Discussion Although gradient methods cannot make large policy changes compared to greedy policy iteration, section 3 implies that these two methods might not be that disparate, since a natural gradient method is moving toward the solution of a policy improvement step. With the overhead of a line search, the methods are even more similar. The benefit is that performance improvement is now guaranteed, unlike in a greedy policy iteration step. It is interesting, and unfortunate, to note that the F does not asymptotically converge to the Hessian, so conjugate gradient methods might be more sensible asymptotically. However, far from the converge point, the Hessian is not necessarily informative, and the natural gradient could be more efficient (as demonstrated in Tetris). The intuition as to why the natural gradient could be efficient far from the maximum, is that it is pushing the policy toward choosing greedy optimal actions. Often, the region (in parameter space) far from from the maximum is where large performance changes could occur. Sufficiently close to the maximum, little performance change occurs (due to the small gradient), so although conjugate methods might converge faster near the maximum, the corresponding performance change might be negligible. More experimental work is necessary to further understand the effectiveness of the natural gradient. Acknowledgments We thank Emo Todorov and Peter Dayan for many helpful discussions. Funding is from the NSF and the Gatsby Charitable Foundation. References [I] S. Amari. Natural gradient works efficiently in learning. 10(2):251- 276, 1998. Neural Computation, [2] J. Baxter and P. Bartlett. Direct gradient-based reinforcement learning. Technical report, Australian National University, Research School of Information Sciences and Engineering, July 1999. [3] D. P. Bertsekas and J. N. Tsitsiklis. Neuro-Dynamic Programming. Athena Scientific, 1996. [4] P. Dayan and G. Hinton. Using em for reinforcement learning. Neural Computation, 9:271- 278 , 1997. [5] S. Kakade. Optimizing average reward using discounted reward. COLT. in press., 200l. [6] V. Konda and J. Tsitsiklis. Actor-critic algorithms. Advances in N eural Information Processing Systems, 12, 2000. [7] D . MacKay. Maximum likelihood and covariant algorithms for independent component analysis. Technical report , University of Cambridge, 1996. [8] P. Marbach and J . Tsitsiklis. Simulation-based optimization of markov reward processes. Technical report, Massachusetts Institute of Technology, 1998. [9] R. Sutton, D. McAllester, S. Singh, and Y. Mansour. Policy gradient methods for reinforcement learning with function approximation. Neural Information Processing Systems, 13, 2000. [10] L. Xu and M. 1. Jordan. On convergence properties of the EM algorithm for gaussian mixtures. Neural Computation, 8(1):129- 151, 1996.
2073 |@word version:1 q7f:3 bf:1 simulation:4 seek:1 tried:1 solid:2 reduction:1 initial:4 substitution:1 score:1 must:1 informative:1 lqg:1 asymptote:1 plot:3 drop:2 update:3 v:4 stationary:6 greedy:14 alone:1 parameterization:3 plane:1 steepest:5 provides:3 sigmoidal:2 height:5 direct:2 become:1 overhead:2 manner:1 expected:2 planning:1 growing:1 nor:1 discounted:1 decreasing:1 little:2 window:2 increasing:1 underlying:3 notation:3 maximizes:1 rmax:1 minimizes:4 finding:1 guarantee:1 every:2 stateaction:1 ro:4 scaled:2 uk:3 control:2 unit:4 bertsekas:1 positive:4 negligible:1 before:2 local:3 engineering:1 tends:2 sutton:2 id:1 becoming:1 abuse:2 might:6 suggests:1 challenging:3 gone:1 unique:1 acknowledgment:1 definite:4 procedure:1 attain:2 argmaxa:2 cannot:2 close:2 writing:1 www:1 equivalent:2 deterministic:1 demonstrated:1 straightforward:4 regardless:1 starting:1 ergodic:1 rule:7 pull:1 lso:1 coordinate:4 anyway:1 lima:1 lsl:1 exact:3 programming:1 us:1 approximated:1 particularly:2 bottom:2 capture:1 parameterize:1 region:1 decrease:1 ran:1 intuition:1 reward:16 dynamic:2 singh:1 solving:1 technically:1 unsealed:1 forced:1 effective:1 london:1 q7r:4 choosing:5 amari:5 timescale:1 online:1 obviously:1 hoc:1 ucl:2 vtj:2 product:1 fr:1 loop:5 sticking:1 description:1 convergence:3 undiscounted:1 p:1 converges:1 oo:1 ac:2 avail:1 ij:1 school:1 sa:15 implies:2 australian:1 direction:6 stochastic:1 mcallester:1 sand:1 argued:2 ao:1 sufficiently:2 cramer:1 s2x:1 exp:9 early:2 estimation:1 clearly:2 gaussian:1 rather:4 improvement:17 likelihood:1 attains:1 sense:3 colon:1 helpful:1 dependent:1 dayan:2 lj:5 initially:1 relation:1 issue:1 among:1 colt:1 logn:3 mackay:3 equal:1 never:1 sampling:2 identical:1 represents:2 blx:1 look:1 future:1 report:3 duplicate:1 few:1 national:1 geometry:2 argmax:3 ourselves:1 phase:2 attempt:1 interest:1 severe:2 replicates:1 introduces:1 mixture:1 extreme:1 arrives:1 tj:12 held:1 wc1n:1 accurate:1 tuple:1 neglecting:1 necessary:1 theoretical:1 increased:2 column:3 rao:1 ar:2 queen:1 cost:2 examining:1 aw:1 chooses:1 st:4 ie:13 squared:5 choose:2 d8:5 derivative:2 return:1 rescaling:2 lookup:1 blind:3 later:1 analyze:1 reached:1 start:2 complicated:1 minimize:2 square:1 ir:1 variance:1 efficiently:1 trajectory:2 plateau:2 reach:2 definition:5 failure:1 proof:3 massachusetts:1 fiat:1 carefully:2 sophisticated:1 e7f:2 higher:1 though:3 just:2 crudely:2 until:1 hand:3 defines:2 olog:1 scientific:1 mdp:5 effect:2 true:1 adjacent:1 game:5 self:6 demonstrate:2 fj:1 tro:1 funding:1 rl:1 numerically:1 cambridge:1 marbach:1 moving:5 actor:3 impressive:1 surface:1 glo:1 add:1 etc:1 curvature:3 showed:1 optimizing:1 dimensionally:1 p7r:2 somewhat:1 converge:4 dashed:2 signal:1 semi:1 relates:1 july:1 sham:2 technical:3 faster:5 characterized:1 af:2 cross:1 sphere:1 controlled:1 neuro:1 expectation:1 metric:13 iteration:10 i09:2 decreased:1 annealing:1 singular:2 source:3 biased:1 ot:2 rest:1 unlike:2 tend:1 legend:1 inconsistent:1 effectiveness:1 jordan:1 near:2 baxter:1 variety:1 todorov:1 fit:1 gave:1 inner:1 intensive:1 whether:1 bartlett:1 f:2 peter:1 speaking:2 hessian:9 action:22 clear:3 detailed:1 locally:1 http:1 sl:1 nsf:1 notice:3 neuroscience:1 prevent:1 asymptotically:5 parameterized:4 family:4 separation:3 decision:2 bound:1 guaranteed:3 distinguish:1 quadratic:1 occur:1 constraint:1 flat:1 regulator:1 speed:1 robustifying:1 extremely:1 poor:3 conjugate:4 smaller:2 slightly:1 em:2 wi:1 kakade:2 making:1 intuitively:1 restricted:2 invariant:3 heart:1 equation:3 emo:1 eventually:2 drastic:3 lieu:1 incurring:1 apply:1 appropriate:3 slower:1 top:4 unfortunate:1 pushing:1 konda:1 move:3 question:1 added:1 occurs:1 strategy:1 dependence:2 exhibit:1 gradient:63 distance:3 thank:1 simulated:1 athena:1 sensible:3 degrade:2 manifold:7 toward:9 length:2 providing:1 equivalently:1 difficult:1 unfortunately:2 disastrous:1 disparate:1 policy:49 boltzmann:1 upper:1 markov:2 finite:3 descent:10 curved:1 defining:1 situation:1 thwarted:1 hinton:1 mansour:1 required:1 connection:1 suggested:1 below:1 oj:1 explanation:1 power:1 natural:33 scheme:1 improve:1 technology:1 mdps:4 coupled:1 iength:1 text:4 understanding:1 tangent:2 interesting:3 approximator:7 foundation:1 integrate:1 agent:2 affine:1 consistent:3 charitable:1 critic:3 unscaled:1 translation:2 compatible:5 last:1 tsitsiklis:3 side:3 understand:2 institute:1 taking:1 benefit:1 curve:11 dimension:1 transition:3 reinforcement:3 far:3 approximate:2 keep:3 search:2 evening:2 why:1 table:1 necessarily:8 did:1 motivation:1 noise:1 s2:2 eural:1 xu:1 gatsby:4 lc:2 exponential:5 ib:1 weighting:4 ix:1 theorem:5 down:1 showing:1 b2x:1 decay:1 dominates:1 concern:1 magnitude:1 hole:2 smoothly:1 ijj:1 bo:1 covariant:3 corresponds:1 satisfies:1 chance:1 goal:2 identity:1 fisher:3 change:6 infinite:1 determined:1 uniformly:1 wt:1 degradation:2 gij:1 tetri:6 experimental:1 la:1 l4:1 select:1 uneven:1 arises:1 phenomenon:1 ex:7
1,178
2,074
Contextual Modulation of Target Saliency Antonio Torralba Dept. of Brain and Cognitive Sciences MIT, Cambridge, MA 02139 torralba@ai. mit. edu Abstract The most popular algorithms for object detection require the use of exhaustive spatial and scale search procedures. In such approaches, an object is defined by means of local features. fu this paper we show that including contextual information in object detection procedures provides an efficient way of cutting down the need for exhaustive search. We present results with real images showing that the proposed scheme is able to accurately predict likely object classes, locations and sizes. 1 Introduction Although there is growing evidence of the role of contextual information in human perception [1], research in computational vision is dominated by object-based representations [5,9,10,15]. In real-world scenes, intrinsic object information is often degraded due to occlusion, low contrast, and poor resolution. In such situations, the object recognition problem based on intrinsic object representations is ill-posed. A more comprehensive representation of an object should include contextual information [11,13]: Obj. representatian == {intrisic obj. model, contextual obj. model}. In this representation, an object is defined by 1) a model of the intrinsic properties of the object and 2) a model of the typical contexts in which the object is immersed. Here we show how incorporating contextual models can enhance target object saliency and provide an estimate of its likelihood and intrinsic properties. 2 Target saliency and object likelihood Image information can be partitioned into two sets of features: local features, VL, that are intrinsic to an object, and contextual features, rUe which encode structural properties of the background. In a statistical framework, object detection requires evaluation of the likelihood function (target saliency function): P(O IVL, va) which provides the probability of presence of the object 0 given a set of local and contextual measurements. 0 is the set of parameters that define an object immersed in a scene: 0 == {on, x, y, i} with on==object class, (x,y)==location in image coordinates and bobject appearance parameters. By applying Bayes rule we can write: P(O IVL, va) = P(vL11 va) P(VL 10, va)P(O Iva) (1) Those three factors provide a simplified framework for representing three levels of attention guidance when looking for a target: The normalization factor, 1/P(VL I va), does not depend on the target or task constraints, and therefore is a bottom-up factor. It provides a measure of how unlikely it is to find a set of local measurements VL within the context va. We can define local saliency as S(x,y) == l/P(vL(x,y) Iva). Saliency is large for unlikely features ina' scene. The second factor, P(VL 10, va), gives the likelihood of the local measurements VL when the object is present at such location in a particular context. We can write P(VL 10, va) ~ P(VL 10), which is a convenient approximation when the aspect of the target object is fully determined by the parameters given by the description O. This factor represents the top-down knowledge of the target? appearance and how it contributes to the search. Regions of the image with features unlikely to belong to the target object are vetoed. and regions with attended features are enhanced. The third factor, the PDF P(O I va), provides context-based priors on object class, location and scale. It is of capital importance for insuring reliable inferences in situations where the local image measurements VL produce ambiguous interpretations. This factor does not depend on local measurements and target models [8,13]. Therefore, the term P(O Iva) modulates the saliency of local image properties when looking for an object of the class On. Contextual priors become more evident if we apply Bayes rule successively in order to split the PDF P( 0 I va) into three factors that model three kinds of context priming on object search: (2) According to this decomposition of the PDF, the contextual modulation of target saliency is a function of three main factors: Object likelihood: P(on Iva) provides the probability of presence of the object class in the scene. If P( On Iva) is very small, then object search need not be initiated (we do not need to look for cars in a living room). On Contextual control of focus of attention: P(x, y I On, va)? This PDDF gives the most likely locations for the presence of object On given context information, and it allocates computational resources into relevant scene regions. Contextual selection of local target appearance: P(tl.va, on). This gives the likely (prototypical) shapes (point of views, size, aspect ratio, object aspect) of the object On in the context Va- Here t == {a, p}, with a==scale and p==aspect ratio. Other parameters describing the appearance of an object in an image can be added. The image features most commonly used for describing local structures are the energy outputs of oriented band-pass filters, as they have been shown to be relevant for the task of object detection [9,10] and scene recognition [2,4,8,12]~ Therefore, the local image representation at the spatial location (x) is given by the vector VL(X) == {v(X,k)}k==l,N with: (3) 1 1 ",.-..... 1 ",.-..... '0 ;> u ;> -a 0 u ;> -a 0 '0: -a 0 P: o1 2 3 4 0: o1 2 3 4 2 3 4 Figure 1: Contextual object prImIng of four objects categories (I-people, 2furniture, 3-vehicles and 4-trees) where i(x) is the input image and gk(X) are oriented band-pass filters defined by gk(i) == e-llxI12/u~e27fj<f~,x>. In such a representation [8], v(i,k) is the output magnitude- at the location i of a complex Gabor filter tuned to the spatial frequency f~. The variable k indexes filters tuned to different spatial frequencies and orientations. On the other ,hand, contextual features have to summarize the structure of the whole image. It has been shown that a holistic low-dimensional encoding of the local image features conveys enough information for a semantic categorization of the scene/context [8] and can be used for contextual priming in object recognition tasks [13]. Such a representation can be achieved by decomposing the image features into the basis functions provided by PCA: an == L L v{x, k) 1/ln{x, k) x k N v(x, k) ~ L an1/ln(x, k) (4) n=l We propose to use the decomposition coefficients vc == {a n }n=l,N as context features. The functions 1/ln are the eigenfunctions of the covariance operator given by v(x, k). By using only a reduced set of components (N == 60 for the rest of the paper), the coefficients {a n }n=l,N encode the main spectral characteristics of the scene with a coarse description of their spatial arrangement. In essence, {a n }n=l,N is a holistic representation as all the regions of the image contribute to all the coefficients, and objects are not encoded individually [8]. In the rest of the paper we show the efficacy of this set of features in context modeling for object detection tasks. 3 Contextual object priming The PDF P( On Iva) gives the probability of presence of the object class On given contextual information. In other words, the PDF P{on Ive) evaluates the consistency of the object On with the context vc. For instance, a car has a high probability of presence in a highway scene but it is inconsistent with an indoor environment. The goal of P(on Ive) is to cut down the number of possible object categories to deal with before- expending computational resources in the object recognition process. The learning of the PDF P(on Ive) == P(ve IOn)P(on)/p(ve) with p(vo) == P(vc IOn)P{on) + P(vc l-,on)P(-,on) is done by approximating the in-class and out-of-class PDFs by a mixture of Gaussians: L P(ve IOn) == L bi,nG(VC;Vi,n, Vi,n) i=l (5) Figure 2: Contextual control of focus of attention when the algorithm is looking for cars (upper row) or heads (bottom row). The model parameters (bi,n, Vi,n, Vi,n) for the object class On are obtained using the EM algorithm [3]. The learning requires the use of few Gaussian clusters (L == 2 provides very good performances). For the learning, the system is trained with a set of examples manually annotated with the .presence/absence of four objects categories (i-people, 2-furniture, 3-vehicles and 4-trees). Fig. 1 shows some typical results from the priming model on the four superordinate categories of objects defined. Note. that the probability function P(on Ive) provides information about the probable presence of one object without scanning the picture. If P( On Ive) > 1th then we can predict that the target is present. On the other hand, if P( On Ive) < th we can predict that the object is likely to be absent before exploring the image. The number of scenes in which the system may be able to take high confidence decisions will depend on different factors such as: the strength of the relationship between the target object and its context and the ability of ve for efficiently characterizing the context. Figure 1 shows some typical results from the priming model for a set of super-ordinate categories of objects. When forcing the model to take binary decisions in all the images (by selecting an acceptance threshold of th == 0.5) the presence/absence of the objects was correctly predicted by the model on 81 % of the scenes of the test set. For each object category, high confidence predictions (th == .1) were made in at least 50% of the tested scene pictures and the presence/absence of each object class was correctly predicted by the model on 95% of those images. Therefore, for those images, we do not need to use local image analysis to decide about the presence/absence of the object. 4 Contextual control of focus of attention One of the strategies that biological visual systems use to deal with the analysis of real-world scenes is to focus attention (and, therefore, computational resources) onto the important image regions while neglecting others. Current computational models of visual attention (saliency maps anQ target detection) rely exclusively on local information or intrinsic object models [6,7,9,14,16]. The control of the focus of attention by contextual information that we propose. here is both task driven (looking for object on) and context driven (given global context information: ve). However, it does riot include any model of the target object at this stage. In our framework, the problem of contextual control of the focus of attention involves the S?? 10 .~ ... 1 ~ CARS o ??? P; ~ _ - filii : ?? fIlIIe?': .. ?? \ \.: ~ .\. tI':,._.: ?? , ??-=- ?? .- 0: ? S: .~ I ? ? "'0 , -: ? ? fill ":I':?.? 1 CARS. ~ t. ,.,:-.,,, ? E 1~ , Q.) ~ .~.~. "'0 ~ 100 HEADS ???? 11 1.8 ?? ~ Q.) 100 ] ~ ? ?\ II Real scale 10 pixels 100 0.4 1 ,---~_"""""""-----R_eal_sc_al--..Je 1 10 pixels 100 tre. ? .tto ? .: 0.4 oReal pose 1 Figure 3: Estimation results of object scale and pose based on contextual features. evaluation of the PDF P(xlon,vo). For the learning, the joint PDF is modeled as a sum of gaussian clusters. Each cluster is decomposed into the product of two gaussians modeling respectively the distribution of object locations and the distribution of contextual features for each cluster: L P(x, vol on) == L bi,n G(x; Xi,n, Xi,n)G(VO; Vi,n, Vi,n) (6) i==l The training set used for the learning of the PDF P(x, vol on) is a subset of'the pictures that contain the object On. The training data is {Vt}t==l,Nt and {Xt}t==l,Nt where Vt are the contextual features of the picture t of the training set and Xt is the location of object On in the image. The model parameters are obtained using the EM algorithm [3,13]. We used 1200 pictures for training and a separate set of 1200 pictures for testing. The success of the PDF in narrowing the region of the focus of attention will depend on the consistency of the relationship between the object and the context. Fig. 2 shows several examples of images and the selected regions based on contextual features when looking for cars and faces. From the PDF P(x, Vo IOn) we selected the region with the highest probability (33% of the image size on average). 87% of the heads present in the test pictures were inside the selected regions. 5 Contextual selection of object appearance models One major problem for computational approaches to object detection is the large variability in object appearance. The classical solution is to explore the space of possible shapes looking for the best match. The main sources of variability in object appearance are size, pose and intra-class shape variability (deformations, style, etc.). We show here that including contextual information can reduce at le.ast the first two sources of variability. For instance, the expected size of people in an image differs greatly between an indoor environment and a perspective view of a street. Both environments produce different patterns of contextual features vo [8]. For the second factor, pose, in the case of cars, there is a strong relationship between the possible orientations of the object and the scene configuration. For instance, looking down a highway, we expect to see the back of the cars, however, in a street view, looking towards the buildings, lateral views of cars are more likely. The expected scale and pose of the target object can be estimated by a regression procedure. The training database used for building the regression is a set of 1000 images in which the target object On is present. For each training image the target Figure 4: Selection of prototypical object appearances based on contextual cues. object was selected by cropping a rectangular window. For faces and cars we define the u == scale as the height of the selected window and the P == pose as the ratio between the horizontal and vertical dimensions of the window (~y/ ~x). On average, this definition of pose provides a good estimation of the orientation for cars but not for heads. Here we used regression using a mixture of gaussians for estimating the conditional PDFs between scale, pose and contextual features: P(u I Va, on) and PCP I va, on). This yields the next regression procedures [3]: (j == Ei Ui,nbi,n G (Va; Vi,n, Vi,n) Ei bi,nG(vO; Vi,n, Vi,n) _ EiPi,nbi,nG(VO;Vi,n, Vi,n) P == Ei bi,nG(VC;Vi,n, Vi,n) (7) The results summarized in fig. 3 show that context is a strong cue for scale selection for the face detection task but less important for the car detection task. On the other hand, context introduces strong constraints on the prototypical point of views of cars but not at all for heads. Once the two parameters (pose and scale) have been estimated, we can build a prototypical model of the target object. In the case of a view-based object representation, the model of the object will consist of a collection of templates that correspond to the possible aspects of the target. For each image the system produces a collection of views, selected among a database of target examples that have the scale and pose given by eqs. (7). Fig. 4 shows some results from this procedure. In the statistical framework, the object detection requires the evaluation of the function P(VL 10, va). We can approximate Input image (target = cars) Object priming and Contextual control Target model selection of focus of attention Integration of local features Target saliency 1 Figure 5: Schematic layout of the model for object detection (here cars) by integration of contextual and local information. The bottom example is an error in detection due to incorrect context identification. P(VL 10, va) ~ P(VL IOn' (J", p). Fig. 5 and 6 show the complete chain of operations and some detection results using a simple correlation technique between the image and the generated object models (100 exemplars) at only one scale. The last image of each row shows the total object likelihood obtained by multiplying the object saliency maps (obtained by the correlation) and the contextual control of the focus of attention. The result shows how the use of context helps reduce false alarms. This results in good detection performances despite the simplicity of the matching procedure used. 6 Conclusion The contextual schema of a scene provides the likelihood of presence, typical locations and appearances of objects within the scene. We have proposed a model for incorporating such contextual cues in the task of object detection. The main aspects of our approach are: 1) Progressive reduction of the window of focus of attention: the system reduces the size of the focus of attention by first integrating contextual information and then local information. 2) Inhibition of target like patterns that are in inconsistent locations. 3) Faster detection of correctly scaled targets that have a pose in agreement with the context. 4) No requirement of parsing a scene into individual objects. Furthermore, once one object has been detected, it can introduce new contextual information for analyzing the rest of the scene. Acknowledglllents The author wishes to thank Dr. Pawan Sinha, Dr. Aude Oliva and Prof. Whitman Richards for fruitful discussions. References [1] Biederman, I., Mezzanotte, R.J., & Rabinowitz, J.C. (1982). Scene perception: detecting and judging objects undergoing relational violations. Cognitive Psychology, 14:143177. Feature maps \ I V t---HXJ---+l .. .. ~ Figure 6: Schema for object detection (e.g. cars) integrating local and giobal information. [2] Carson, C., Belongie, S., Greenspan, H., and Malik, J. (1997). Region-based image querying. Proc. IEEE W. on Content-Based Access of Image and Video Libraries, pp: 42-49. [3] Gershnfeld, N. The nature of mathematical modeling. Cambridge university press, 1999. [4] Gorkani, M. M., Picard, R. W. (1994). Texture orientation for sorting photos 'at a glance'. Proc. Int. Conf. Pat. Rec., Jerusalem, Vol. I: 459-464. [5] Heisle, B., T. Serre, S. Mukherjee and T. Poggio. (2001) Feature Reduction and Hierarchy of Classifiers for Fast Object Detection in Video Images. In: Proceedings of 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, IEEE Computer Society Press, Jauai, Hawaii. [6] Itti, L., Koch, C., & Niebur, E. (1998). A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Analysis and Machine Vision, 20(11):1254. [7] Moghaddam, B., & Pentland, A. (1997). Probabilistic Visual Learning for Object Representation. IEEE Trans. Pattern Analysis and Machine Vision, 19(7):696-710. [8] Oliva, A., & Torralba, A. (2001). Modeling the Shape of the Scene: A holistic representation of the spatial envelope. Int. Journal of Computer Vision, 42(3):145-175. [9] Rao, R.P.N., Zelinsky, G.J., Hayhoe, M.M., & Ballard, D.H. (1996). Modeling saccadic targeting in visual search. NIPS 8. Cambridge, MA: MIT Press. [10] Schiele, B., Crowley, J. L. (2000) Recognition without Correspondence using Multidimensional Receptive Field Histograms, Int. Journal of Computer Vision, Vol. 36(1):31-50. [11] Strat, T. M., & Fischler, M. A. (1991). Context-based vision: recognizing objects using information from both 2-D and 3-D imagery. IEEE trans. on Pattern Analysis and Machine Intelligence, 13(10): 1050-1065. [12] Szummer, M., and Picard, R. W. (1998). Indoor-outdoor image classification. In IEEE intl. workshop on Content-based Access of Image and Video Databases, 1998. [13] Torralba, A., & Sinha, P. (2001). Statistical context priming for object detection. IEEE Proc. Of Int. Conf in Compo Vision. [14] Treisman, A., & Gelade, G. (1980). A feature integration theory of attention. Cognitive Psychology, Vol. 12:97-136. [15] Viola, P. and Jones, M. (2001). Rapid object detection using a boosted cascade of simple features. In: Proceedings of 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2001), IEEE Computer Society Press, Jauai, Hawaii. [16] Wolfe, J. M. (1994). Guided search 2.0. A revised model of visual search. Psychonomic Bulletin and Review, 1:202-228
2074 |@word decomposition:2 covariance:1 attended:1 reduction:2 configuration:1 efficacy:1 selecting:1 exclusively:1 tuned:2 current:1 contextual:37 nt:2 pcp:1 parsing:1 shape:4 cue:3 selected:6 intelligence:1 compo:1 coarse:1 provides:9 contribute:1 location:11 detecting:1 height:1 mathematical:1 become:1 incorrect:1 inside:1 introduce:1 expected:2 rapid:2 growing:1 brain:1 decomposed:1 window:4 provided:1 estimating:1 kind:1 iva:6 multidimensional:1 ti:1 scaled:1 classifier:1 control:7 intrisic:1 before:2 local:19 despite:1 encoding:1 analyzing:1 initiated:1 modulation:2 bi:5 testing:1 differs:1 procedure:6 gabor:1 cascade:1 convenient:1 matching:1 word:1 confidence:2 integrating:2 onto:1 targeting:1 selection:5 operator:1 context:23 applying:1 ast:1 fruitful:1 map:3 layout:1 attention:15 jerusalem:1 rectangular:1 resolution:1 simplicity:1 rule:2 crowley:1 fill:1 coordinate:1 target:27 enhanced:1 hierarchy:1 agreement:1 wolfe:1 recognition:7 rec:1 richards:1 cut:1 mukherjee:1 database:3 bottom:3 role:1 narrowing:1 region:10 nbi:2 highest:1 environment:3 ui:1 schiele:1 fischler:1 trained:1 depend:4 basis:1 whitman:1 joint:1 fast:1 detected:1 exhaustive:2 encoded:1 posed:1 ive:6 cvpr:1 ability:1 propose:2 product:1 relevant:2 holistic:3 description:2 cluster:4 requirement:1 cropping:1 produce:3 categorization:1 intl:1 object:91 help:1 pose:11 exemplar:1 eq:1 strong:3 predicted:2 involves:1 guided:1 annotated:1 filter:4 vc:6 human:1 require:1 probable:1 biological:1 exploring:1 koch:1 predict:3 major:1 torralba:4 estimation:2 proc:3 individually:1 highway:2 mit:3 gaussian:2 super:1 greenspan:1 boosted:1 encode:2 focus:11 pdfs:2 likelihood:7 ivl:2 contrast:1 greatly:1 inference:1 vl:14 unlikely:3 pixel:2 among:1 ill:1 orientation:4 classification:1 spatial:6 integration:3 field:1 once:2 ng:4 manually:1 represents:1 progressive:1 look:1 jones:1 others:1 few:1 oriented:2 ve:5 comprehensive:1 individual:1 occlusion:1 pawan:1 detection:20 acceptance:1 intra:1 picard:2 evaluation:3 introduces:1 mixture:2 violation:1 chain:1 moghaddam:1 fu:1 neglecting:1 poggio:1 allocates:1 tree:2 guidance:1 deformation:1 riot:1 sinha:2 instance:3 modeling:5 rao:1 subset:1 recognizing:1 scanning:1 probabilistic:1 enhance:1 treisman:1 imagery:1 successively:1 zelinsky:1 dr:2 hawaii:2 cognitive:3 conf:2 style:1 itti:1 summarized:1 coefficient:3 int:4 vi:14 vehicle:2 view:7 schema:2 bayes:2 degraded:1 characteristic:1 efficiently:1 yield:1 saliency:12 correspond:1 identification:1 accurately:1 niebur:1 multiplying:1 definition:1 evaluates:1 energy:1 frequency:2 pp:1 conveys:1 popular:1 knowledge:1 car:16 back:1 strat:1 done:1 furthermore:1 stage:1 correlation:2 hand:3 tre:1 horizontal:1 ei:3 glance:1 rabinowitz:1 aude:1 building:2 serre:1 contain:1 semantic:1 deal:2 ambiguous:1 essence:1 carson:1 pdf:11 evident:1 complete:1 vo:7 image:35 psychonomic:1 belong:1 interpretation:1 measurement:5 cambridge:3 ai:1 consistency:2 access:2 inhibition:1 etc:1 perspective:1 driven:2 forcing:1 binary:1 success:1 vt:2 living:1 ii:1 expending:1 reduces:1 match:1 faster:1 va:18 schematic:1 prediction:1 regression:4 oliva:2 vision:9 histogram:1 normalization:1 achieved:1 ion:5 background:1 source:2 envelope:1 rest:3 eigenfunctions:1 inconsistent:2 obj:3 structural:1 presence:11 split:1 enough:1 psychology:2 reduce:2 absent:1 pca:1 antonio:1 band:2 category:6 reduced:1 judging:1 estimated:2 correctly:3 write:2 vol:5 four:3 threshold:1 capital:1 immersed:2 sum:1 decide:1 decision:2 furniture:2 correspondence:1 tto:1 strength:1 constraint:2 scene:21 dominated:1 aspect:6 according:1 poor:1 em:2 partitioned:1 ln:3 resource:3 describing:2 photo:1 decomposing:1 gaussians:3 operation:1 apply:1 spectral:1 top:1 include:2 build:1 prof:1 approximating:1 classical:1 society:4 malik:1 added:1 arrangement:1 strategy:1 saccadic:1 receptive:1 separate:1 thank:1 lateral:1 street:2 anq:1 o1:2 index:1 relationship:3 modeled:1 ratio:3 gk:2 upper:1 vertical:1 revised:1 pentland:1 pat:1 situation:2 relational:1 looking:8 head:5 variability:4 viola:1 hxj:1 biederman:1 ordinate:1 nip:1 trans:3 gelade:1 able:2 hayhoe:1 superordinate:1 perception:2 pattern:7 indoor:3 summarize:1 including:2 reliable:1 video:3 rely:1 representing:1 scheme:1 vetoed:1 library:1 picture:7 prior:2 review:1 ina:1 fully:1 expect:1 prototypical:4 querying:1 row:3 last:1 template:1 characterizing:1 face:3 bulletin:1 dimension:1 world:2 author:1 insuring:1 commonly:1 made:1 simplified:1 collection:2 approximate:1 cutting:1 global:1 belongie:1 xi:2 search:8 an1:1 nature:1 ballard:1 contributes:1 priming:8 complex:1 rue:1 main:4 whole:1 alarm:1 fig:5 je:1 tl:1 wish:1 outdoor:1 third:1 down:4 xt:2 showing:1 undergoing:1 evidence:1 intrinsic:6 incorporating:2 consist:1 false:1 workshop:1 importance:1 modulates:1 texture:1 magnitude:1 sorting:1 likely:5 appearance:9 explore:1 visual:6 ma:2 conditional:1 goal:1 towards:1 room:1 absence:4 content:2 typical:4 determined:1 total:1 pas:2 people:3 szummer:1 dept:1 tested:1
1,179
2,075
Modeling the Modulatory Effect of Attention on Human Spatial Vision Laurent Itti Computer Science Department, Hedco Neuroscience Building HNB-30A, University of Southern California, Los Angeles, CA 90089-2520, U.S.A. J oehen Braun nstitute of Neuroscience and School of Computing, University of Plymouth, Plymouth Devon PL4 8AA, U.K. Christof Koch Computation and Neural Systems Program, MC 139-74, California Institute of Technology, Pasadena, CA 91125 , U.S.A. Abstract We present new simulation results , in which a computational model of interacting visual neurons simultaneously predicts the modulation of spatial vision thresholds by focal visual attention, for five dual-task human psychophysics experiments. This new study complements our previous findings that attention activates a winnertake-all competition among early visual neurons within one cortical hypercolumn. This "intensified competition" hypothesis assumed that attention equally affects all neurons, and yielded two singleunit predictions: an increase in gain and a sharpening of tuning with attention. While both effects have been separately observed in electrophysiology, no single-unit study has yet shown them simultaneously. Hence, we here explore whether our model could still predict our data if attention might only modulate neuronal gain, but do so non-uniformly across neurons and tasks. Specifically, we investigate whether modulating the gain of only the neurons that are loudest, best-tuned, or most informative about the stimulus, or of all neurons equally but in a task-dependent manner, may account for the data. We find that none of these hypotheses yields predictions as plausible as the intensified competition hypothesis, hence providing additional support for our original findings. 1 INTRODUCTION Psychophysical studies as well as introspection indicate that we are not blind outside the focus of attention, and that we can perform simple judgments on objects not being attended to [1], though those judgments are less accurate than in the presence of attention [2, 3]. While attention thus appears not to be mandatory for early vision, there is mounting experimental evidence from single-neuron electrophysiology [4, 5, 6, 7, 8, 9, 10], human psychophysics [11 , 12, 13, 14,3, 2, 15, 16] and human functional imaging experiments [17, 18, 19, 20, 21, 22, 23] that focal visual attention modulates, top-down, activity in early sensory processing areas. In the visual domain, this modulation can be either spatially-defined (i.e., neuronal activity only at the retinotopic location attended to is modulated) or feature-based (i.e., neurons with stimulus preference matching the stimulus attended to are enhanced throughout the visual field), or a combination of both [7, 10, 24]. Computationally, the modulatory effect of attention has been described as enhanced gain [8, 10], biased [4] or intensified [14, 2] competition, enhanced spatial resolution [3], sharpened neuronal tuning [5, 25] or as modulated background activity [19], effective stimulus strength [26] or noise [15]. One theoretical difficulty in trying to understand the modulatory effect of attention in computational terms is that, although attention profoundly alters visual perception, it is not equally important to all aspects of vision. While electrophysiology demonstrates "increased firing rates" with attention for a given task, psychophysics show "improved discrimination thresholds" on some other tasks, and functional magnetic resonance imaging (fMRI) reports "increased activation" for yet other tasks, the computational mechanism at the origin of these observations remains largely unknown and controversial. While most existing theories are associated to a specific body of data, and a specific experimental task used to engage attention in a given experiment, we have recently proposed a unified computational account [2] that spans five such tasks (32 thresholds under two attentional conditions, i.e., 64 datapoints in total). This theory predicts that attention activates a winner-take-all competition among neurons tuned to different orientations within a single hyper column in primary visual cortex (area VI). It is rooted in new information-theoretic advances [27], which allowed us to quantitatively relate single-unit activity in a computational model to human psychophysical thresholds. A consequence of our "intensified competition hypothesis" is that attention both increases the gain of early visual neurons (by a factor 3.3), and sharpens their tuning for the orientation (by 40%) and spatial frequency (by 30%). While gain modulation has been observed in some of the single-unit studies mentioned above [8, 10] (although much smaller effects are typically reported, on the order of 10-15%, probably because these studies do not use dual-task paradigms and thus poorly engage the attention of the animal towards or away from the stimulus of interest), and tuning modulation has been observed in other single-unit studies [5, 25], both gain and tuning modulation have not been simultaneously observed in a single electrophysiological set of experiments [10]. In the present study, we thus investigate alternatives to our intensified competition hypothesis which only involve gain modulation. Our previous results [2] have shown that both increased gain and sharper tuning were necessary to simultaneously account for our five pattern discrimination tasks, if those modulatory effects were to equally affect all visual neurons at the location of the stimulus and to be equal for all tasks. Thus, we here extend our computational search space under two new hypotheses: First, we investigate whether attention might only modulate the gain of selected sub-populations of neurons (responding the loudest, best tuned , or most informative about the stimulus) in a task-independent manner. Second, we investigate whether attention might equally modulate the gain of all visual neurons responding to the stimulus, but in a task-dependent manner. Thus, the goal of the present study is to determine, using new computational simulations, whether the modulatory effect of attention on early visual processing might be explained by gain-only modulations, if such modulations are allowed to be sufficiently complex (affecting only select visual neurons , or task-dependent). Although attention certainly affects most stages of visual processing, we here continue to focus on early vision, as it is widely justified by electrophysiological and fMRI evidence that some modulation does happen very early in the processing hierarchy [5, 8, 9, 23]. 2 PSYCHOPHYSICAL DATA Our recent study [2] measured psychophysical thresholds for three pattern discrimination tasks (contrast, orientation and spatial frequency discriminations), and two spatial masking tasks (32 thresholds) . We used a dual-task paradigm to measure thresholds either when attention was fully available to the task of interest (presented in the near periphery), or when it was poorly available because engaged elsewhere by a concurrent attention-demanding task (a letter discrimination task at the center of the display). The results are summarized in Fig. 1 and [2]. 'I' c:: A B 0 4:: c:: .g .S .g ~ E .;: .Q... .S ~ .? ~ c:: .S .~ .~ ~ ... 8 Q '" E c:: ~ - 10.1 W ~ [' ? Fully Attended 0 Poorly Attended I t ; :~ f- 10-2 ~ c:: CU ::s : t~ I ~ ; a4:: !! /1 ",0.3 Q) > oJ ~ 0.1 00 ~ 10'2 10" Mask contrast 0 ~ 1:3 0.2 0- ?- I C ~ .... II I Q c:: .g .sc:: II I 0.2 8 .~ 0.4 0.6 Contrast 0.8 1 .~ 0 6 I ~ 4j 2 00 ! 0.2 0.4 0.6 0.8 1 Contrast 'I' D E I't I:J\ I:J\ .S .S .lC '" .lC 0.4 ~ ... :E 1;; l!! 0.3 '" C8 0.2 E c:: W ~ f-~ 0.1 0 h ? Fully Attended o Poorly Attended 0.4 ~ ... 1;; 0 " c:: W I 20 40 I 60 80 Mask e - Target e C) I I '" l!! 0.3 '" c E 1+ 0 '" :E ~ f-~ 0.2 0.1 0 0.5 Mask w / Target w 2 Figure 1: Psychophysical data from Lee et ai. Central targets appeared at 0 - 0.8? eccentricity and measured 0.4? across. Peripheral targets appeared for 250 ms at 4? eccentricity, in a circular aperture of 1.5 0 ? They were either sinusoidal gratings (B, C) or vertical stripes whose luminance profile was given by the 6th derivative of a Gaussian (A, D, E) . Mask patterns were generated by superimposing 100 Gabor filters , positioned randomly within the circular aperture (A, D, E). Thresholds were established with an adaptive staircase method (80 trials per block). A complex pattern of effects is observed, with a strong modulation of orientation and spatial frequency discriminat ions (B, C) , smaller modulation of contrast discriminations (A) , and modulation of contrast masking that depends on stimulus configurations (D, E). These complex observations can be simultaneously accounted for by our computational model of one hypercolumn in primary visual cortex. 3 COMPUTATIONAL MODEL The model developed to quantitatively account for this data comprises three successive stages [14, 27]. In the first stage, a bank of Gabor-like linear filters (12 orientations and 5 spatial scales) analyzes a given visual location, similarly to a cortical hyper column. In the second stage, filters nonlinearly interact through both a selfexcitation component, and a divisive inhibition component that is derived from a pool of similarly-tuned units. With E)."o being the linear response from a unit tuned to spatial period A and orientation (), the response R)."o after interactions is given by (see [27] for additional details): Linear filters Divisive inhibition R Decision ).,,0 - (S)O L + +B (A.E)."o) ' W)."o(A',()') (A.E)..!,O')O (1) ' ()..! ,O') EA x 8 where: W (A' ()') = )., ,0 , exp (_ (log(A') -log(A))2 _ (()' - ())2) 2A2 2A2 )., (2) 0 is a 2D Gaussian weighting function centered around (A, ()) whose widths are determined by the scalars Ao and A).,. The neurons are assumed to be noisy, with noise variance V{o given by a generalized Poisson model: V{o = (3(R)."o + <:). The third stage relates activity in the population of interacting noisy units to behavioral discrimination performance. To allow us to quantitatively predict thresholds from neural activity for any task, our decision stage assumes that observers perform close to an unbiased efficient statistic, that is, the best possible estimator (in the statistical estimation sense) of the characteristics of the stimulus given the noisy neuronal responses. This methodology (described further in [27]) allows us to quantitatively compute thresholds in any behavioral situation, and eliminates the need for task-dependent assumptions about the decision strategy used by the observers. 4 RESULTS and DISCUSSION The 10 free model parameters (Fig. 2) were automatically adjusted to best fit the psychophysical data from all experiments, using a multidimensional downhill simplex with simulated annealing overhead (see [27]) , running on our 16CPU Linux Beowulf system (16 x 733 MHz, 4 GB RAM, 0.5 TB disk; see http://iLab . usc. edu/beo/). Parameters were simultaneously adjusted for both attentional conditions; that is, the total fit error was the sum of the error obtained with the baseline set of parameters on the poorly attended data, and of the error obtained with the same parameters plus some attentional perturbation on the fully attended data. Thus, no bias was given to any of the two attentional conditions. For the "separate fits" (Fig. 2), all parameters were allowed to differ with attention [2], while only the interaction parameters b, 8) could differ in the "intensified competition" case. The "loudest filter" was the one responding loudest to the entire visual pattern presented (stimulus + mask), the "best-tuned filter" was that responding best to the stimulus component alone, and the "most informative filter" was that for which the Fisher information about the stimulus was highest (see Model Parameters y,o: S, ?: 11 S: y - Interaction strength ~: ~~; ~::~:~ ~~~~g Attentional manipulations Top-down Attention Dark noise Light noise Semisaturation ~ Ul r Q (" ) :!: ,,(oct) B/ A S/A " ,,(oct) (I) 11 en ? / Rmu, -.- .--- 0.01 0.11 - very good fit overall - all parameters biologically plausible - attention significantly modulates interactions and noise ~ BI A ***** Top-down Attention Parameters (I) m 0.65 0.01 Parameters 7.03 0.24 3.00 1.5e-9 8635.52 ? c - - -........./ 0.31 3~:!~ ~:: I: 9(") I: .. (OCI) BIA S IA ~ EI R""" Parameters - Poorly ~:~~~ \~"\ ..~ '\, 9 (" ) o .. (OCI) ... QQQ ... 'y".... - Fully 1.45 ~05 Top-down Attention . - no modulation of contrast detection threshold - no modulation of orientation thresholds - no modulation of period thresholds - poor prediction of masking - filter tuning too narrow - gain modulation too large - no modulation of orientation th resholds - no modulation of period thresholds - contrast discrimination and masking well fit - only fit predicting broad pooling in spatial period - noise parameters unrealistic **** Parameters [affects all filters, but differently for each task] .~ I ~. ,,,,,,. V-"\. 14.71 ~~ Q~Q (I) Poorly 16:: '\\ [stimulus-dependent; only affects filter most infonnative a~out target stimulus] 1-8. * Fully 1.40 ~~ Top-down Attention c \.'\ 1.09 0.21 12.16 ~~ Q~~ (I) ~ \'""i. 6.62 1 ~:~ [stimulus-dependent; only affects filter beslluned to/arget stimulus] != UlLL Poorly 1.03 ~. Q ~~ ... Ul" Fully 2.22 ~56 [stimulus-dependent; only affects filter responding most to whole (targr+maSk) stimulus] Top-down Attention ~(I) 0.96 5.39 0.56 0.02 S/A P :::l= OLL -I c - very good fit overall - all parameters biologically plausible - modulation of orientation thresholds slightly underestimated - contrast masking with variable mask orientation not perfectly predicted - 8. 92 0.41 20.80 0.31 EI R""" ",S - ***** Poorly 3.941.40 3.551.00 0 0 (0) O,,(OCI) I: 9 (") I: ,,(oct) (I) - ... CU 1. 10 8.05 0.17 0.03 Fully co. ,SE c 0 -0 :::l 0.30 10.12 Parameters - (1).- (I) 1.85 - ***** Top-down Attention "5 .- "c Poorly 3.78 3.421 .80 9.6913.19 0.440.36 23.01 23.90 0.810.18 0 0 (0) Q.LL Ul Fully Parameters - (I) CU:!:: Discussion -..........- (1" 9( ") (I" .,(oct) I: 9(") r .,(oct) BIA S/A , ?/R""", 3.4<-4 0.01 * Fully - no contrast discrimination "dipper" - power-law rather than sigmoidal contrast response (S=O) - modulation of orientation thresholds slightly underestimated - noise parameter unrealistic *** Poorl y - very good fit overall - gain modulation unrealistically high , especially for orientation discrimination (filter gain when attending to orientation is > 20x poorly attended) Figure 2: Attentional modulation hypot heses and corresponding model parameters. See next page for the corresponding model predictions on our five tasks, for the hypot heses shown. The middle column shows which parameters were allowed to differ with attention, and t he best-fit values for both attentional conditions. Contrast Increment Spatial Frequency Discrimination Orientation Discrimination Contrast Masking, Variable Mask e R lM~ ,.~ C ! 0.2 ~ .0.1 810 " ~" 10 -, 0.2 0.4 0.6 Contrast 0.8 1 R02 I . ~ 0.1 ??? "----'--,--"'7----l o 10 .2 ?0 10 " 0.2 Mask contrast **** - Fully attended Poorly attended 10"2 10" Mask contrast 0.2 0.4 0.6 Contrast 0.8 1 ***** 0.4 0.6 Contrast C ? 0.8 - Fully attended Poorly attended 8'0" 0.4 0.6 0.8 1 !:~~ 0.2 0.4 0.6 Contrast 0.8 Poorly attended o 0.2 0.4 0.6 f ? ? Fullyattended POOrly attended 0.8 1 c:B I %4 '.. 10'> 00 10-' 0.2 0.4 0.6 Contrast 0.8 1 00 0.2 0.4 0.6 Contrast 1 **** 0; ? l" - Fully attended Poorly attended ?10" o f.3~ r i j ~ 01 '. ? ~ 00 ** 4 . 0.4 0.6 Contrast 0.8 1 0.3 ~ ~ ~0.1 . ~ l" 10'> 10.2 10" Mask contrast *** 00 00 0.2 0.4 0.6 Contrast 0.8 ***** 0.2 0.4 0.6 Contrast 0.8 1 1 8~ t 6 i, 2 00 I 0.4 0.6 Contrast ? 0.5 1 MaskwfTarget ***** 2 **** 0.4[ 8 ~ 03 , 8 0.2 '. I ? ***** ***** ? Fully attended Poorly attended + (J) o I 0.8 2 t 0020406080 Mask O- TargetOr} I o 0.5 1 MaskwfTarget + 2 (J) 0.4[ 8 t 0.3 0.2 ~ ' ? ? 0.2 2 !? 0.5 1 MaskwlTargetw $ 0.1 ? *** !o,\~ ~ ~'0" 0.5 1 MaskwlTargetw Qj 2 . I 0.4~ I . o ?020406080 Mask O- TargetOr} ? Fully attended Poorly attended ***** ~ ~ 0.2 . 0. 3 8 02 $ 0.1 10 ~ 10" 10 " Mask contrast j t Qj ***** R0.2 0.1 * ? Fully attended Poorly attended ? 0.8 2 0.4~ 0.2 $" 0.1 60 80 Mask6 - Target6(O) ? 2 0.5 1 MaskwlTargetw **** t:0.4~ I . I 10~ Mask contrast + 0.3 o ?020406080 *** Contrast f3~ 1!! ~ .~ ~ I I 1 ? Mask6 - Target6(O) Contrast .~ ~ I ~ 0.1 !!'- 0.2 *** ?0 L ***** ? Fullyattended ? **** R0.2 ~ ??? ?0 \ I ' 0.2 o ?020406080 Mask6 - Target6(O) **** I 6 2 1 0.4[ 8 0.3 ~ ~ *** ~ ?0 ***** ~ 8 C ~ 0.1 lM~ ,.~I ~ ? Fully attended Poorly attended 2 ?0 **** ! I 6 Contrast Masking, Variable Mask 00 ? ? + ~ 0.1 1 0020406080 MaskO - TargetO(O} o ***** ***** 0.5 1 MaskwlTargetw 2 Figure 3: Model predictions for t he different attentional modulation hypotheses studied. The different rows correspond to t he different attentional manipulations studied, as labeled in t he previous figure. Ratings (stars below t he plots) were derived from t he residual error of t he fits . [14, 27]). Finally, in the "task-dependent" case, the gain of all filters was affected equally (parameter ')'), but with three different values for the contrast (discrimination and masking), orientation and spatial frequency tasks. Overall, very good fits were obtained in the "separate fits" and "intensified competition" conditions (as previously reported) , as well as in the "most informative filter" and "taskdependent" conditions (Fig. 3) , while the two remaining hypotheses yielded very poor predictions of orientation and spatial frequency discriminations. In the "most informative filter" case, the dipper in the contrast increment thresholds was missing because the nonlinear response function of the neurons converged to a power law rather than the usually observed sigmoid [27]; thus, this hypothesis lost some of its appeal because of its lower biological plausibility. More importantly, a careful analysis of the very promising results for the "task-dependent" case also revealed their low biological plausibility, with a gain modulation in excess of 20-fold being necessary to explain the orientation discrimination data (Fig. 2). In summary, we found that none of the simpler (gain only) attentional manipulations studied here could explain as well the psychophysical data as our previous manipulation, "intensified competition," which implied that attention both increases the gain and sharpens the tuning of early visual neurons. Two of the four new manipulations studied yielded good quantitative model predictions: affecting the gain of the filter most informative about the target stimulus, and affecting the gain of all filters in a task-dependent manner. In both cases, however, some of the internal model parameters associated with the fits were biologically unrealistic, thus reducing the plausibility of these two hypotheses. In all manipulations studied, the greatest difficulty was in trying to account for the orientation and spatial frequency discrimination data without unrealistically high gain changes (greater than 20-fold). Our results hence provide additional evidence for the hypothesis that sharpening of tuning may be necessary to account for these thresholds, as was originally suggested by our separate fits and our intensified competition hypothesis and has been recently supported by new investigations [16]. Acknowledgements This research was supported by the National Eye Institute, the National Science Foundation, the NSF-supported ERC center at Caltech, the National Institutes for Mental Health, and startup funds from the Charles Lee Powell Foundation and the USC School of Engineering. References [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] Braun J & Sagi D. P ercept Psychophys , 1990;48(1):45- 58. Lee DK , Itti L, Koch C et al. Nat Neurosci, 1999;2(4):375-81. Yeshurun Y & Carrasco M. Nature, 1998;396(6706) :72- 75 . Moran J & Desimone R . Science , 1985 ;229(4715) :782- 4. Spitzer H, Desimone R & Moran J. Sci ence, 1988 ;240(4850):338- 40 . Chelazzi L, Miller EK, Duncan J et al. Nature , 1993;363(6427):345- 7. Motter BC. J Neurosci, 1994;14(4):2178-89. Treue S & Maunsell JH. Nature, 1996;382(6591):539- 41. Luck SJ, Chelazzi L, Hillyard SA et al. J Neurophysiol, 1997;77(1) :24- 42. Treue S & Trujillo JCM . Nature, 1999 ;399(6736) :575- 579 . Nakayama K & Mackeben M. Vision Res, 1989 ;29(11) :1631- 47. Bonnel AM , Stein JF & Bertucci P. Q J Exp Psychol A, 1992 ;44(4):601- 26. Lee DK , Koch C & Braun J. Vision R es, 1997;37(17):2409- 18 . [14] Itti L, Braun J, Lee DK et al. In NIPS*ll. MIT Press, 1999; pp. 789- 795. [15] Dosher BA & Lu ZL. Vision Res, 2000;40(10-12):1269- 1292. [16] Carrasco M, Penpeci-Talgar C & Eckstein M. Vision Res, 2000;40(10-12):1203- 1215. [17] Corbett a M, Miezin FM, Dobmeyer S et al. Science, 1990;248(4962):1556- 9. [18] Rees G, Frackowiak R & Frith C. Science, 1997;215(5301):835- 8. [19] Chawla D, Rees G & Friston KJ. Nat Neurosci, 1999;2(7):671- 676. [20] Brefczynski JA & DeYoe EA. Nat Neurosci, 1999;2(4):370- 374. [21] Corbetta M, Kincade JM, Ollinger JM et al. Nat Neurosci, 2000;3(3):292- 297. [22] Kanwisher N & Wojciulik E. Nat Rev Neurosci, 2000;1:91- 100. [23] Ress D, Backus BT & Heeger DJ. Nat Neurosci, 2000;3(9):940- 945. [24] Barcelo F, Suwazono S & Knight RT. Nat Neurosci, 2000;3(4) :399- 403. [25] Desimone R & Duncan J . Annu Rev Neurosci, 1995 ;18:193- 222 . [26] Reynolds JH, Pasternak T & Desimone R. Neuron, 2000;26(3):703- 714 . [27] Itti L, Koch C & Braun J. J Opt Soc Am A, 2000;11(11):1899- 1917.
2075 |@word trial:1 cu:3 middle:1 sharpens:2 disk:1 simulation:2 attended:26 configuration:1 tuned:6 bc:1 reynolds:1 existing:1 activation:1 yet:2 happen:1 informative:6 plot:1 fund:1 mounting:1 discrimination:16 alone:1 selected:1 mental:1 location:3 preference:1 successive:1 sigmoidal:1 simpler:1 five:4 overhead:1 behavioral:2 manner:4 kanwisher:1 mask:16 automatically:1 cpu:1 jm:2 retinotopic:1 spitzer:1 developed:1 unified:1 finding:2 sharpening:2 quantitative:1 multidimensional:1 braun:5 demonstrates:1 zl:1 unit:7 maunsell:1 christof:1 engineering:1 sagi:1 consequence:1 laurent:1 firing:1 modulation:24 might:4 plus:1 studied:5 co:1 bi:1 lost:1 block:1 powell:1 area:2 gabor:2 significantly:1 matching:1 close:1 center:2 missing:1 attention:34 resolution:1 estimator:1 attending:1 importantly:1 datapoints:1 population:2 increment:2 enhanced:3 hierarchy:1 target:6 engage:2 hypothesis:12 origin:1 qqq:1 stripe:1 carrasco:2 predicts:2 labeled:1 observed:6 highest:1 luck:1 backus:1 knight:1 mentioned:1 arget:1 neurophysiol:1 yeshurun:1 differently:1 frackowiak:1 effective:1 sc:1 startup:1 hyper:2 outside:1 whose:2 widely:1 plausible:3 statistic:1 noisy:3 interaction:4 poorly:20 competition:11 los:1 r02:1 eccentricity:2 object:1 measured:2 school:2 sa:1 strong:1 soc:1 grating:1 predicted:1 indicate:1 differ:3 filter:18 centered:1 human:5 ja:1 ao:1 discriminat:1 investigation:1 opt:1 biological:2 adjusted:2 koch:4 sufficiently:1 around:1 exp:2 predict:2 lm:2 early:8 a2:2 estimation:1 modulating:1 concurrent:1 mit:1 activates:2 gaussian:2 rather:2 treue:2 derived:2 focus:2 contrast:33 baseline:1 sense:1 am:2 dependent:10 typically:1 entire:1 bt:1 pasadena:1 overall:4 dual:3 among:2 orientation:18 resonance:1 spatial:14 animal:1 psychophysics:3 field:1 equal:1 f3:1 broad:1 introspection:1 fmri:2 report:1 stimulus:20 quantitatively:4 simplex:1 randomly:1 simultaneously:6 national:3 loudest:4 usc:2 detection:1 interest:2 investigate:4 circular:2 certainly:1 light:1 accurate:1 desimone:4 plymouth:2 necessary:3 re:3 theoretical:1 increased:3 column:3 modeling:1 ence:1 mhz:1 infonnative:1 too:2 reported:2 rees:2 lee:5 pool:1 linux:1 sharpened:1 central:1 bonnel:1 ek:1 derivative:1 itti:4 account:6 sinusoidal:1 star:1 summarized:1 blind:1 vi:1 depends:1 observer:2 masking:8 variance:1 largely:1 characteristic:1 miller:1 yield:1 judgment:2 correspond:1 mc:1 none:2 lu:1 converged:1 explain:2 frequency:7 pp:1 associated:2 gain:22 electrophysiological:2 singleunit:1 positioned:1 ea:2 appears:1 originally:1 methodology:1 response:5 improved:1 though:1 stage:6 ei:2 nonlinear:1 building:1 effect:8 staircase:1 unbiased:1 hence:3 spatially:1 ll:2 width:1 rooted:1 m:1 generalized:1 trying:2 theoretic:1 recently:2 charles:1 sigmoid:1 functional:2 ilab:1 winner:1 extend:1 he:9 trujillo:1 ai:1 tuning:9 focal:2 similarly:2 erc:1 winnertake:1 dj:1 hillyard:1 cortex:2 inhibition:2 recent:1 mandatory:1 periphery:1 manipulation:6 continue:1 caltech:1 dosher:1 analyzes:1 additional:3 greater:1 r0:2 determine:1 paradigm:2 period:4 ii:2 relates:1 plausibility:3 equally:6 target6:3 prediction:7 vision:9 poisson:1 ion:1 justified:1 background:1 affecting:3 separately:1 unrealistically:2 annealing:1 underestimated:2 biased:1 eliminates:1 probably:1 pooling:1 dipper:2 near:1 presence:1 revealed:1 affect:7 fit:14 perfectly:1 fm:1 angeles:1 qj:2 whether:5 gb:1 ul:3 modulatory:5 se:1 involve:1 dark:1 stein:1 http:1 nsf:1 alters:1 neuroscience:2 per:1 profoundly:1 affected:1 motter:1 four:1 threshold:18 luminance:1 imaging:2 ram:1 corbetta:1 sum:1 bia:2 letter:1 throughout:1 decision:3 duncan:2 display:1 fold:2 chelazzi:2 yielded:3 activity:6 strength:2 aspect:1 span:1 c8:1 department:1 peripheral:1 combination:1 poor:2 across:2 smaller:2 slightly:2 deyoe:1 rev:2 biologically:3 explained:1 hypot:2 computationally:1 remains:1 previously:1 mechanism:1 available:2 away:1 magnetic:1 chawla:1 pasternak:1 alternative:1 original:1 top:7 responding:5 assumes:1 running:1 remaining:1 a4:1 res:1 especially:1 psychophysical:7 implied:1 strategy:1 primary:2 rt:1 southern:1 attentional:10 separate:3 simulated:1 intensified:9 sci:1 providing:1 sharper:1 relate:1 ba:1 unknown:1 perform:2 vertical:1 neuron:18 observation:2 situation:1 interacting:2 perturbation:1 rating:1 complement:1 nonlinearly:1 hypercolumn:2 eckstein:1 devon:1 california:2 narrow:1 established:1 nip:1 suggested:1 below:1 perception:1 pattern:5 usually:1 psychophys:1 appeared:2 program:1 tb:1 oj:1 ia:1 unrealistic:3 demanding:1 difficulty:2 power:2 greatest:1 predicting:1 friston:1 residual:1 technology:1 eye:1 pl4:1 psychol:1 health:1 kj:1 acknowledgement:1 law:2 fully:17 oci:3 foundation:2 controversial:1 bank:1 row:1 jcm:1 elsewhere:1 summary:1 accounted:1 supported:3 free:1 bias:1 allow:1 understand:1 jh:2 institute:3 cortical:2 sensory:1 adaptive:1 excess:1 sj:1 aperture:2 assumed:2 corbett:1 search:1 promising:1 nature:4 ca:2 nakayama:1 frith:1 interact:1 complex:3 domain:1 neurosci:9 whole:1 noise:7 profile:1 allowed:4 body:1 neuronal:4 fig:5 en:1 lc:2 sub:1 comprises:1 downhill:1 heeger:1 weighting:1 third:1 down:7 annu:1 miezin:1 specific:2 moran:2 appeal:1 dk:3 evidence:3 oll:1 modulates:2 taskdependent:1 nat:7 electrophysiology:3 explore:1 visual:18 beo:1 scalar:1 aa:1 oct:5 modulate:3 goal:1 careful:1 towards:1 jf:1 fisher:1 change:1 specifically:1 determined:1 uniformly:1 reducing:1 total:2 engaged:1 experimental:2 hnb:1 superimposing:1 divisive:2 e:1 select:1 internal:1 support:1 modulated:2
1,180
2,076
Exact differential equation population dynamics for Integrate-and-Fire neurons Julian Eggert * HONDA R&D Europe (Deutschland) GmbH Future Technology Research Carl-Legien-StraBe 30 63073 Offenbach/Main, Germany julian. eggert@hre-ftr.f.rd.honda.co.jp Berthold Bauml Institut fur Robotik und Mechatronik Deutsches Zentrum fur Luft und Raumfahrt (DLR) oberpfaffenhofen Berthold.Baeuml@dlr.de Abstract Mesoscopical, mathematical descriptions of dynamics of populations of spiking neurons are getting increasingly important for the understanding of large-scale processes in the brain using simulations. In our previous work, integral equation formulations for population dynamics have been derived for a special type of spiking neurons. For Integrate- and- Fire type neurons , these formulations were only approximately correct. Here, we derive a mathematically compact, exact population dynamics formulation for Integrate- and- Fire type neurons. It can be shown quantitatively in simulations that the numerical correspondence with microscopically modeled neuronal populations is excellent. 1 Introduction and motivation The goal of the population dynamics approach is to model the time course of the collective activity of entire populations of functionally and dynamically similar neurons in a compact way, using a higher descriptionallevel than that of single neurons and spikes. The usual observable at the level of neuronal populations is the populationaveraged instantaneous firing rate A(t), with A(t)6.t being the number of neurons in the population that release a spike in an interval [t, t+6.t). Population dynamics are formulated in such a way, that they match quantitatively the time course of a given A(t), either gained experimentally or by microscopical, detailed simulation. At least three main reasons can be formulated which underline the importance of the population dynamics approach for computational neuroscience. First, it enables the simulation of extensive networks involving a massive number of neurons and connections, which is typically the case when dealing with biologically realistic functional models that go beyond the single neuron level. Second, it increases the analytical understanding of large-scale neuronal dynamics , opening the way towards better control and predictive capabilities when dealing with large networks. Third, it enables a systematic embedding of the numerous neuronal models operating at different descriptional scales into a generalized theoretic framework, explaining the relationships, dependencies and derivations of the respective models. Early efforts on population dynamics approaches date back as early as 1972, to the work of Wilson and Cowan [8] and Knight [4], which laid the basis for all current population-averaged graded-response models (see e.g. [6] for modeling work using these models). More recently, population-based approaches for spiking neurons were developed, mainly by Gerstner [3, 2] and Knight [5]. In our own previous work [1], we have developed a theoretical framework which enables to systematize and simulate a wide range of models for population-based dynamics. It was shown that the equations of the framework produce results that agree quantitatively well with detailed simulations using spiking neurons, so that they can be used for realistic simulations involving networks with large numbers of spiking neurons. Nevertheless, for neuronal populations composed of Integrate-and-Fire (I&F) neurons, this framework was only correct in an approximation. In this paper, we derive the exact population dynamics formulation for I&F neurons. This is achieved by reducing the I&F population dynamics to a point process and by taking advantage of the particular properties of I&F neurons. 2 2.1 Background: Integrate-and-Fire dynamics Differential form We start with the standard Integrate- and- Fire (I&F) model in form of the wellknown differential equation [7] (1) which describes the dynamics of the membrane potential Vi of a neuron i that is modeled as a single compartment with RC circuit characteristics. The membrane relaxation time is in this case T = RC with R being the membrane resistance and C the membrane capacitance. The resting potential v R est is the stationary potential that is approached in the no-input case. The input arriving from other neurons is described in form of a current ji. In addition to eq. (1), which describes the integrate part of the I&F model, the neuronal dynamics are completed by a nonlinear step. Every time the membrane potential Vi reaches a fixed threshold () from below, Vi is lowered by a fixed amount Ll > 0, and from the new value of the membrane potential integration according to eq. (1) starts again. if Vi(t) = () (from below) . (2) At the same time, it is said that the release of a spike occurred (i.e., the neuron fired), and the time ti = t of this singular event is stored. Here ti indicates the time of the most recent spike. Storing all the last firing times , we gain the sequence of spikes {t{} (spike ordering index j, neuronal index i). 2.2 Integral form Now we look at the single neuron in a neuronal compound. We assume that the input current contribution ji from presynaptic spiking neurons can be described using the presynaptic spike times tf, a response-function ~ and a connection weight W? . ',J ji(t) = Wi ,j ~(t - tf) (3) l: l: j f Integrating the I&F equation (1) beginning at the last spiking time tT, which determines the initial condition by Vi(ti) = vi(ti - 0) - 6., where vi(ti - 0) is the membrane potential just before the neuron spikes, we get 1 Vi(t) = v Rest + fj(t - t:) + l: Wi ,j l: a(t - t:; t - tf) , j f - Vi(t:)) e- S / T (4) with the refractory function fj(s) = - (v Rest (5) and the alpha-function r ds" JSI_ S Sf a(s; s') = e-[sf -S"J/T ~(s") . (6) If we start the integration at the time ti* of the spike before the last spike, then for ti* :::; t < ti the membrane potential is given by an expression like eq. (4), where ti is replaced by t:i* . Especially we can now express v( ti - 0) and therefore the initial condition for an integration starting at tT in terms of ti* and v(ti* - 0). In this way, we can proceed repetitively and move back into the past. After some simple algebra this results in Vi(t) = v Rest + l:ry(t-t{)+ l:Wi,j l:a(oo ;t - tf) , f ~ vfef(t) (7) j f ~-------y~------~ v~yn(t) with a refractory function wich differs in the scale factor from that in eq. (5) ry(s) = -6. e- S / T ? (8) The components vref(t) and v?n(t) to the membrane potential indicate refractory and synaptic components to the neuron i, respectively, as normally used in the Spike- Response- Model (SRM) notation [2]. Both equations (4) and (7) formulate the neuronal dynamics using a refractory component, which depends on the own spike releases of a neuron, and a synaptic component, which comprises the integrated input contribution to the membrane potential by arrival of spikes from other neurons 2. The synaptic component is based on the alpha-function characteristic of isolated arriving spikes, with an increase of the membrane potential after spike arrival and a subsequent exponential decrease. 1 Strictly speaking, the constants vRest, T, () and ,6, and the function 1]( s) may vary for each neuron, so that they should be written with a subindex i [similarly for n(s; s') , which may vary for each connection j -+ i so that we should write it with subindices i, j]. For the sake of clarity, we omit these indices here . 2S0 the I&F model can be formulated as a special case of the Spike- Response- Model, which defines the neuronal dynamics in the integral formulation. The comparison of the equivalent expressions eq. (4) and eq. (7) reveals an interesting property of the I&F model. They look very similar, but in eq. (4), the refractory component depends only on the time elapsed since the last spike (thus reflecting a renewal property, sometimes also called a short term memory for refractory properties), whereas in eq. (7), it depends on a sum of the contributions of all past spikes. The simpler form of the refractory contribution in eq. (4) is achieved at the cost of an alpha-function that now depends on the time t - ti elapsed since the last own spike in addition to the times t - tf elapsed since the release of spikes at the presynaptic neurons j that provide input to i. In eq. (7) , we have a more complex refractory contribution, but an alpha-function that does not depend on the last own spike time any more. 2.3 Probabilistic spike release Probabilistic firing is introduced into the I&F model eq. (4) resp. (7) by using threshold noise. The spike release of each neuron is controlled by a hazard function >.(v), so that >.(v)dt = Prob. that a neuron with membrane potential v spikes in [t , t + dt) (9) When a neuron spikes, we proceed as usual: The membrane potential is reset by a fixed amount 6. and the I&F dynamics continues. Population dynamics 3 3.1 Density description Descriptions of neuronal populations usually assume a neuronal density function p(t; X) which depends on the state variables X of the neurons. The density quantifies the likelihood that a neuron picked out of the population will be found in the vicinity of the point X in state space, p(t; X) dX = Portion of neurons at time t with state in [X, X + dX) (10) If we know p(t; X) , the population activity A(t) can be easily calculated. Using the hazard function >'(t; X), the instantaneous population activity (spikes per time) can be calculated by computing the spike release averaged over the population, A(t) = J dX >.(t; X) p(t; X) (11) The population dynamics is then given by the time course of the neuronal density function p(t; X), which changes because each neuron evolves according to its own internal dynamics, e.g. after a spike release and the subsequent reset of the membrane potential. The main challenge for the formulation of a population dynamics resides in selecting a low-dimensional state space [for an easy calculation of A(t)] and a suitable form for gtp(t; X). As an example, for the population dynamics for I&F neurons it would be straightforward to use the membrane potential v from eq. (1) as the state variable X. But this leads to a complicated density dynamics, because the dynamics for v(t) consist of a continuous (differential equation (1)) and a discrete part (spike generation). Therefore, here we concentrate on an alternative description that allows a compact formulation of the desired I&F density dynamics. 3.2 Exact population dynamics for I&F neurons Which is the best state space for a population dynamics of I&F neurons? For the formulation of a population dynamics, it is usually assumed that the synaptic contributions to the membrane potential are identical for all neurons. This is the case if we group all neurons of the same dynamical type and with identical connectivity patterns into one population. That is, we say that neurons i and i' belong to the same population if Wi,j = Wi',j for all j (for simulations of realistic networks of spiking neurons, this will of course never be exactly the case, but it is reasonable to assume that a grouping of neurons into populations can be achieved to a good approximation) . In our case, looking at eq. (4), we see that , since o:(s, s') depends on s = t - ti and therefore on the own last spike time, the synaptic contribution to the membrane potential differs according to the state of the neuron. Thus we regard eq. (7). Here, we see that for identical connectivity patterns Wi,j, the synaptic contributions are the same for all neurons, because 0:(00, s') does not depend on the own spike time any more. Which are then the state variables of eq. (1) for the density description? We see that, for a fixed synaptic contribution, the membrane potential Vi is fully determined by the set of the own past spiking times {tf}. But this would mean an infinite-dimensional density for the state description of a population, and, accordingly, a computationally overly expensive calculation of the population activity A(t) according to eq. (11). To avoid this we take advantage of a particular property of the I&F model. According to eq. (8), the single spike refractory contributions 'TJ(s) are exponential. Since any sum of exponential functions with common relaxation constant T can be again expressed as as an exponential function with the same T , we can write instead of vrf(t) from eq. (7) (12) Now the membrane potential Vi(t) only depends on the time of the last own spike ti and the refractory contribution amplitude modulation factor at the last spike 'TJi . That is, we have transferred the effect of all spikes previous to the last one into 'TJi. In addition, we have to care about updating of ti and 'TJi when a neuron spikes so that we get 3 'TJi --+ 'TJi = 1 + 'TJie -(t-tn!T , (13) ti --+ ti = t . The effect of taking into account more than the most recent spike ti in the refractory component vief(t) leads to a modulation factor 'TJi greater than 1, in particular if spikes come in a rapid succession so that refractory contributions can accumulate. Instead of using a modulation factor 'TJi the effect of previous spikes can also be taken into account by introducing an effective last spiking time ii. (14) vi"f(t) = 'TJ(t - in = 'TJi'TJ(t - tn , where ii and 'TJi are connected by i; = t; + TIn'TJi (15) The effect of i* is sort of funny. Because of 'TJi ::::: 1 it holds for the effective last spiking time ii ::::: ti. This means, that , while at a given time t it is allways ti :::; t, it happens that ii ::::: t, meaning the neurons behave as if they would spike in the future. 3Here, the order of reemplacement matters; first we have to reemplace 1]:, then ti. For the membrane potential we get now instead of eq. (7) Vi(t) = v Rest + ry(t - tn + 2..: Wi ,j 2..: 0:(00; t - t;) (16) f j and for the update rule for the effective last spiking time *) ' * = f (t 'tA tAi* - +tA i i with t; follows (17) (18) Therefore we can regard the dimensionality of the state space of the I&F dynamics as 1-dimensional in the description of eq. (16). The dynamics of the single I&F neurons now turns out to be very simple: Calculate the membrane potential Vi(t) using eq. (16) together with the state variable t;, and check if Vi(t) exceeds the threshold. If not, move forward in time and calculate again. If the membrane potential exceeds threshold, update according to eq. (17) and then proceed with the calculation of Vi(t) as normal. t; Using this single neuron dynamics , we can now proceed to gain a population dynamics using a density p(t; t*). The time t is here the explicit time dependence, whereas t* denote the state variable of the population. By fixing t* and the synaptic contribution vsyn(t) to the membrane potential, the state of a neuron is fully determined and the hazard function can be written as ,X[vsyn(t); t*]. The dynamics of the density p(t; t*) is then calculated as follows. Changes of p(t; t*) occur when neurons spike and t* is updated according to eq. (17). The hazard function controls the spike release, and, therefore, the change of p(t; t*). For chosen state variables, p(t; t*) decreases due to spiking of the neurons with the fixed t*, and increases because neurons with other t'* spike and get updated in just that way that after updating their state variable falls around t*. This occurs according to the reemplacement rule eq. (17) when f(t, t'*) = t* . (19) Taking all together the dynamics of the density p(t; t*) is given by decrease due to same state t* spiking A -ftp(t;t*) + 1 = '-,X[vsyn(t); t*]p(t; t*)' (20) + 00 -00 dt'* 8[J(t, t'*) - t*] ,X[vsyn(t); t'*] p(t; t'*) increase due to spiking of neurons with other states t'* The population activity can then be calculated using the density according to eq. (11) as follows 1 +00 A(t) = - 00 dt* ,X[vsyn(t); t*] p(t; t*) (21) Remark that the expression for the density dynamics (eq. 20) automatically conserves the norm of the density, so that 1 +00 - 00 dt* p( t ; t*) = const , (22) which is a necessary condition because the number of neurons participating in the dynamics must remain constant. 4 Simulations The dynamics of a population of I&F neurons , represented by the time course of their joint activity, can now be easily calculated in terms of the differential equation (20) , if the neuronal state density of the neuronal population p(t; i*) and the synaptic input vsyn(t) are known. This means that all we have to store is the density p(t; i*) for past and future effective last spiking times i* 4 . Favorably for numerical simulations, only a limited time window of i* around the actual time t is needed for the dynamics. The activity A(t) only appears as an auxiliary variable that is calculated with the help of the neuronal density. In figure 1 the simulation results for populations of of spiking neurons are shown. The neurons are uncoupled and a hazard function A(V) = ~ e2,B(v-e) (23) , TO with spike rate at threshold liTO = 1.0ms- 1 , a kind of inverse temperature (3 = 2.0, which controls the noise level, and the threshold = 1.0. The other parameters of the model in eq. (1) are: resting potential v Rest = 0, jump in membrane potential after spike release ~ = 1 and time constant T = 20ms. This parameters are chosen to be biologicaly plausible. e A (spikes/ms) 0.14 0.12 0.1 0.08 0.06 0.04 r------- 0.02 : b) = II "" n "":: :: !l :~ I 1\ *:\ ! \_ .. ----2 '-1 1l ! \_, .. ----: _______ J: ! . . . . ..'! ~,' j ~ I:, r------- vsyr'i-'_ _1_00_ _15 _0_ _ 20_0 _ _2_50_ _ 30_0 ----,1 (ms) o~ I c) 100 150 200 250 300 I (ms) Figure 1: Activity A(t) of simulated populations of neurons. The neurons are uncoupled and to each neuron the same synaptic field vsyn(t), ploted in c) and d), is applied. a) shows the activity A(t) for a population of I&F neurons simulated on the one hand as N = 10000 single neurons (solid line) using eq. (7) and on the other hand using the density dynamics eq. (20) (dashed line). In b) the activity A(t) of a population ofI&F neurons (dashed line) and a population of SRM neurons with renewal (solid line) are compared. For all simulations the same parameters as specified in the text were used. The simulations show that the density dynamics eq. (20) reproduces the activity A( t) of a population of single I&F neurons almost perfect, with the exception of the noise in the single neuron simulations due to the finite size effects. This holds even for the peaks occuring at the steps of the applied synaptic field v syn (t), although the density dynamics is entirely based on differential equations and one would therefore not expect such an excellent match for fast changes in activity. S ll (t) only appears as a scalar in the dynamics, so that no integration over time takes 4V Y place here. The simulations also show that there can be a big difference between I&F and SRM neurons with renewal. Because of the accumulation of the refractory effects of all former spikes in the case of I&F neurons the activity A(t) is generaly lower than for the SRM neurons with renewal and the higher the absolute actitvity level the bigger is the difference between both. 5 Conclusions In this paper we derived an exact differential equation density dynamics for a population of I&F neurons starting from the microscopical equations for a single neuron. This density dynamics allows a compuationaly efficient simulation of a whole population of neurons. In future work we want to simulate a network of connected neuronal populations. In such a network of populations (indexed e.g. by x) , a self-consistent system of differential equations based on the population's p(x, t; i*) and A(x, t) emerges if we constrain ourselves to neuronal populations connected synaptically according to the constraints given by the pool definition [2]. In this case, two neurons i and j belong to pools x and y, if Wi,j = W(x, y). This allows us to write for the synaptic component of the membrane potential v syn(x,t) = 2: W (x , y) y 1 00 ds'a(oo;s')A(y,t-s') (24) 0 Using the alpha-function a(oo ; s') as introduced in (6), and a "nice" responsefunction ~ for the input current time course after a spike, we can write eq. (24) using differential equations that use A(y, t) as input. This results in a system that is based entirely on differential equations and is very cheap to compute. References [1] J. Eggert and J.L. van Hemmen. Modeling neuronal assemblies: Theory and implementation. N eural Computation, 13(9):1923- 1974, 200l. [2] W. Gerstner. Population dynamics of spiking neurons: Fast transients, asynchronous states and locking. Neural Computation, 12:43- 89 , 2000. [3] W . Gerstner and J . L. van Hemmen. Associative memory in a network of 'spiking' neurons. Network, 3:139- 164, 1992. [4] B. W. Knight . Dynamics of encoding in a populations of neurons. J. Gen. Physiology, 59:734- 766 , 1972. [5] B. W. Knight. Dynamics of Encoding in Neuron Populations: Some General Mathematical Features. Neural Comput., 12:473- 518, 2000. [6] Z. Li. A neural model of contour integration in the primary visual cortex. Neural Comput. , 10(4):903- 940, 1998. [7] H. C. Tuckwell. Introduction to Th eoretical N eurobiology. Cambridge University Press, Cambridge, 1988. [8] H. R. Wilson and J. D. Cowan. Excitatory and inhibitory interactions in localized populations of model neurons. Biophys. J ., 12:1- 24, 1972.
2076 |@word norm:1 underline:1 simulation:15 solid:2 initial:2 selecting:1 past:4 current:4 dx:3 written:2 must:1 realistic:3 subsequent:2 numerical:2 enables:3 cheap:1 update:2 stationary:1 accordingly:1 beginning:1 short:1 honda:2 simpler:1 mathematical:2 rc:2 differential:10 rapid:1 ry:3 brain:1 automatically:1 actual:1 window:1 notation:1 circuit:1 vref:1 kind:1 developed:2 every:1 ti:22 exactly:1 control:3 normally:1 omit:1 yn:1 before:2 encoding:2 firing:3 modulation:3 approximately:1 dynamically:1 co:1 limited:1 range:1 averaged:2 differs:2 physiology:1 integrating:1 get:4 accumulation:1 equivalent:1 go:1 straightforward:1 starting:2 wich:1 formulate:1 vsyn:7 rule:2 embedding:1 population:56 updated:2 resp:1 massive:1 exact:5 carl:1 expensive:1 conserve:1 updating:2 continues:1 vrf:1 calculate:2 connected:3 ordering:1 decrease:3 knight:4 und:2 locking:1 dynamic:48 depend:2 algebra:1 predictive:1 basis:1 easily:2 joint:1 represented:1 derivation:1 fast:2 effective:4 approached:1 plausible:1 say:1 associative:1 advantage:2 sequence:1 analytical:1 interaction:1 reset:2 date:1 gen:1 fired:1 subindices:1 description:7 participating:1 getting:1 produce:1 perfect:1 ftp:1 derive:2 oo:3 help:1 fixing:1 eq:31 auxiliary:1 indicate:1 come:1 concentrate:1 vrest:1 correct:2 tji:11 transient:1 mathematically:1 strictly:1 hold:2 around:2 normal:1 vary:2 early:2 tf:6 avoid:1 wilson:2 derived:2 release:10 fur:2 indicates:1 mainly:1 likelihood:1 check:1 entire:1 typically:1 integrated:1 germany:1 special:2 integration:5 renewal:4 eurobiology:1 field:2 never:1 identical:3 look:2 future:4 quantitatively:3 opening:1 composed:1 zentrum:1 microscopically:1 replaced:1 ourselves:1 fire:6 tj:3 integral:3 necessary:1 respective:1 institut:1 indexed:1 desired:1 isolated:1 theoretical:1 modeling:2 cost:1 introducing:1 srm:4 stored:1 dependency:1 density:22 peak:1 systematic:1 probabilistic:2 pool:2 together:2 connectivity:2 again:3 li:1 account:2 potential:25 de:1 matter:1 vi:17 depends:7 picked:1 portion:1 start:3 sort:1 capability:1 complicated:1 contribution:13 compartment:1 characteristic:2 succession:1 reach:1 synaptic:12 definition:1 e2:1 gain:2 emerges:1 dimensionality:1 amplitude:1 syn:2 back:2 reflecting:1 appears:2 higher:2 dt:5 ta:2 response:4 formulation:8 just:2 d:2 hand:2 nonlinear:1 defines:1 effect:6 former:1 vicinity:1 tuckwell:1 ll:2 self:1 m:5 generalized:1 theoretic:1 tt:2 occuring:1 eggert:3 tn:3 temperature:1 fj:2 meaning:1 instantaneous:2 recently:1 common:1 functional:1 spiking:19 microscopical:2 ji:3 refractory:13 jp:1 belong:2 occurred:1 resting:2 functionally:1 eoretical:1 accumulate:1 cambridge:2 rd:1 similarly:1 lowered:1 europe:1 cortex:1 operating:1 own:9 recent:2 wellknown:1 compound:1 store:1 greater:1 care:1 dashed:2 ii:5 exceeds:2 match:2 calculation:3 repetitively:1 hazard:5 bigger:1 controlled:1 ftr:1 involving:2 sometimes:1 achieved:3 synaptically:1 background:1 addition:3 whereas:2 want:1 interval:1 singular:1 rest:5 cowan:2 easy:1 dlr:2 expression:3 effort:1 resistance:1 proceed:4 speaking:1 remark:1 subindex:1 detailed:2 amount:2 inhibitory:1 neuroscience:1 overly:1 per:1 write:4 discrete:1 express:1 group:1 nevertheless:1 threshold:6 clarity:1 relaxation:2 sum:2 prob:1 inverse:1 place:1 laid:1 reasonable:1 almost:1 funny:1 entirely:2 correspondence:1 activity:13 occur:1 constraint:1 constrain:1 sake:1 simulate:2 transferred:1 according:10 membrane:25 describes:2 remain:1 increasingly:1 wi:8 evolves:1 biologically:1 happens:1 taken:1 computationally:1 equation:14 agree:1 tai:1 turn:1 needed:1 know:1 alternative:1 assembly:1 completed:1 const:1 especially:1 graded:1 move:2 capacitance:1 spike:48 occurs:1 primary:1 dependence:1 usual:2 said:1 simulated:2 presynaptic:3 reason:1 modeled:2 relationship:1 index:3 julian:2 favorably:1 implementation:1 collective:1 neuron:78 finite:1 behave:1 looking:1 introduced:2 specified:1 extensive:1 connection:3 elapsed:3 uncoupled:2 beyond:1 below:2 usually:2 dynamical:1 pattern:2 challenge:1 memory:2 gtp:1 event:1 suitable:1 technology:1 numerous:1 text:1 nice:1 understanding:2 fully:2 expect:1 interesting:1 generation:1 localized:1 integrate:7 consistent:1 s0:1 storing:1 course:6 excitatory:1 last:14 asynchronous:1 arriving:2 explaining:1 wide:1 taking:3 fall:1 absolute:1 van:2 regard:2 berthold:2 calculated:6 resides:1 contour:1 forward:1 jump:1 alpha:5 compact:3 observable:1 dealing:2 reproduces:1 reveals:1 assumed:1 continuous:1 quantifies:1 excellent:2 gerstner:3 complex:1 main:3 motivation:1 noise:3 big:1 arrival:2 whole:1 eural:1 gmbh:1 neuronal:19 hemmen:2 comprises:1 explicit:1 sf:2 exponential:4 comput:2 third:1 tin:1 grouping:1 consist:1 gained:1 importance:1 biophys:1 visual:1 expressed:1 lito:1 scalar:1 determines:1 goal:1 formulated:3 towards:1 experimentally:1 change:4 determined:2 infinite:1 reducing:1 hre:1 called:1 est:1 ofi:1 exception:1 internal:1
1,181
2,077
Constructing Distributed Representations Using Additive Clustering Wheeler Ruml Division of Engineering and Applied Sciences Harvard University 33 Oxford Street, Cambridge, MA 02138 ruml@eecs.harvard.edu Abstract If the promise of computational modeling is to be fully realized in higherlevel cognitive domains such as language processing, principled methods must be developed to construct the semantic representations used in such models. In this paper, we propose the use of an established formalism from mathematical psychology, additive clustering, as a means of automatically constructing binary representations for objects using only pairwise similarity data. However, existing methods for the unsupervised learning of additive clustering models do not scale well to large problems. We present a new algorithm for additive clustering, based on a novel heuristic technique for combinatorial optimization. The algorithm is simpler than previous formulations and makes fewer independence assumptions. Extensive empirical tests on both human and synthetic data suggest that it is more effective than previous methods and that it also scales better to larger problems. By making additive clustering practical, we take a significant step toward scaling connectionist models beyond hand-coded examples. 1 Introduction Many cognitive models posit mental representations based on discrete substructures. Even connectionist models whose processing involves manipulation of real-valued activations typically represent objects as patterns of 0s and 1s across a set of units (Noelle, Cottrell, and Wilms, 1997). Often, individual units are taken to represent specific features of the objects and two representations will share features to the degree to which the two objects are similar. While this arrangement is intuitively appealing, it can be difficult to construct the features to be used in such a model. Using random feature assignments clouds the relationship between the model and the objects it is intended to represent, diminishing the model?s value. As Clouse and Cottrell (1996) point out, hand-crafted representations are tedious to construct and it can be difficult to precisely justify (or even articulate) the principles that guided their design. These difficulties effectively limit the number of objects that can be encoded, constraining modeling efforts to small examples. In this paper, we investigate methods for automatically synthesizing feature-based representations directly from the pairwise object similarities that the model is intended to respect. This automatic Table 1: An 8-feature model derived from consonant confusability data. With c = 0.024, the model accounts for 91.8% of the variance in the data. Wt. Objects with feature Interpretation .350 f? front unvoiced fricatives .243 dg back voiced stops .197 p k unvoiced stops (without t) .182 b v? front voiced .162 ptk unvoiced stops .127 mn nasals .075 dgv?z? z voiced (without b) .049 ptkf?s? s unvoiced approach eliminates the manual burden of selecting and assigning features while providing an explicit design criterion that objectively connects the representations to empirical data. After formalizing the problem, we will review existing algorithms that have been proposed for solving it. We will then investigate a new approach, based on combinatorial optimization. When using a novel heuristic search technique, we find that the new approach, despite its simplicity, performs better than previous algorithms and that, perhaps more important, it maintains its effectiveness on large problems. 1.1 Additive Clustering We will formalize the problem of constructing discrete features from similarity information using the additive clustering model of Shepard and Arabie (1979). In this framework, abbreviated A DCLUS, clusters represent arbitrarily overlapping discrete features. Each of the k features has a non-negative real-valued weight wk , and the similarity between two objects i and j is just the sum of the weights of the features they share. If f ik is 1 if object i has feature k and 0 otherwise, and c is a real-valued constant, then the similarity of i and j is modeled as X s?ij = wk fik fjk + c . k This class of models is very expressive, encompassing non-hierarchical as well as hierarchical arrangements of clusters. An example model, derived using the ewindclus-klb algorithm described below, is shown in Table 1. The representation of each object is simply the binary column specifying its membership or absence in each cluster. Additive clustering is asymmetric in the sense that only the shared features of two objects contribute to their similarity, not the ones they both lack. (This is the more general formulation, as an additional feature containing the set complement of the original feature could always be used to produce such an effect.) With a model formalism in hand, we can then phrase the problem of constructing feature assignments as simply finding the A DCLUS model that best matches the given similarity data using the desired number of features. The fit of a model (comprising F , W , and c) to a matrix, S, can be quantified by the variance accounted for (VAF), which compares the model?s accuracy to merely predicting using the mean similarity: P ?ij )2 i,j (sij ? s VAF = 1 ? P ?)2 i,j (sij ? s A VAF of 0 can always be achieved by setting all wk to 0 and c to s?. 2 Previous Algorithms Additive clustering is a difficult 0-1 quadratic programming problem and only heuristic methods, which do not guarantee an optimal model, have been proposed. Many different approaches have been taken: Subsets: Shepard and Arabie (1979) proposed an early algorithm based on subset analysis that was clearly superseded by Arabie?s later work below. Hojo (1983) also proposed an algorithm along these lines. We will not consider these algorithms further. Non-discrete Approximation: Arabie and Carroll (1980) and Carroll and Arabie (1983) proposed the two-stage indclus algorithm. In the first stage, cluster memberships are treated as real values and optimized for each cluster in turn by gradient descent. At the same time, a penalty term for non-0-1 values is gradually increased. Afterwards, a combinatorial clean-up stage tries all possible changes to 1 or 2 cluster memberships. Experiments reported below use the original code, modified slightly to handle large instances. Random initial configurations were used. Asymmetric Approximation: In the sindclus algorithm, Chaturvedi and Carroll (1994) optimize anP asymmetric model with two sets of cluster memberships, having the form s?ij = k wk fik gjk + c. By considering each cluster in turn, this formulation allows a fast method for determining each of F , G, and w given the other two. In practice, F and G often become identical, yielding an A DCLUS model. Experiments reported below use both a version of the original implementation that has been modified to handle large instances and a reimplemented version (resindclus) that differs in its behavior at boundary cases (handling 0 weights, empty clusters, ties). Models from runs in which F and G did not converge were each converted into several A DCLUS models by taking only F , only G, their intersection, or their union. The weights and constants of each model were optimized using constrained least-squares linear regression (Stark and Parker, 1995), ensuring non-negative cluster weights, and the one with the highest VAF was used. Alternating Clusters: Kiers (1997) proposed an element-wise simplified sindclus algorithm, which we abbreviate as ewindclus. Like sindclus, it considers each cluster in turn, alternating between the weights and the cluster memberships, although only one set of clusters is maintained. Weights are set by a simple regression and memberships are determined by a gradient function that assumes object independence and fixed weights. The experiments reported below use a new implementation, similar to the reimplementation of sindclus. Expectation Maximization: Tenenbaum (1996) reformulated A DCLUS fitting in probabilistic terms as a problem with multiple hidden factorial causes, and proposed a combination of the EM algorithm, Gibbs sampling, and simulated annealing to solve it. The experiments below use a modified version of the original implementation which we will notate as em-indclus. It terminates early if 10 iterations of EM pass without a change in the solution quality. (A comparison with the original code showed this modification to give equivalent results using less running time.) Unfortunately, it is not clear which of these approaches is the best. Most published comparisons of additive clustering algorithms use only a small number of test problems (or only artificial data) and report only the best solution found within an unspecified amount of time. Because the algorithms use random starting configurations and often return solutions of widely varying quality even when run repeatedly on the same problem, this leaves it unclear which algorithm gives the best results on a typical run. Furthermore, different Table 2: The performance of several previously proposed algorithms on data sets from psychological experiments. indclus sindclus re-sindclus ewindclus Name VAF IQR VAF IQR r VAF IQR r VAF IQR r animals-s 77 75?80 66 65?65 8 78 79 ?80 12 64 60?69 4 numbers 83 81?86 84 82 ?86 5 78 75?81 7 82 79?85 5 workers 83 82?85 81 79?83 9 84 82?85 7 67 63?72 2 consonants 89 89?90 88 87?89 6 81 80?82 5 51 44?57 1 animals 71 69?74 66 66?66 9 66 66?66 13 72 71 ?73 26 80 80?80 78 78?79 7 68 65?72 5 74 73?75 17 letters Table 3: The performance of indclus and em-indclus on the human data sets. indclus em-indclus Name n k VAF IQR r VAF IQR animals-s 10 3 80 80?80 23 80 80?80 numbers 10 8 91 90?91 157 90 89?90 89 88?89 89 87 87?89 workers 14 7 consonants 16 8 91 91?91 291 91 91?91 71 69?74 1 N/A animals 26 12 letters 30 5 82 82?83 486 82 82?83 algorithms require very different running times, and multiple runs of a fast algorithm with high variance in solution quality may produce a better result in the same time as a single run of a more predictable algorithm. The next section reports on a new empirical comparison that addresses these concerns. 2.1 Evaluation of Previous Algorithms We compared indclus, both implementations of sindclus, ewindclus, and emindclus on 3 sets of problems. The first set is a collection of 6 typical data sets from psychological experiments that have been used in previous additive clustering work (originally by Shepard and Arabie (1979), except for animals-s, Mechelen and Storms (1995), and animals, Chaturvedi and Carroll (1994)). The number of objects (n) and the number of features used (k) are listed for each instance as part of Table 3. The second set of problems contains noiseless synthetic data derived from A DCLUS models with 8, 16, 32, 64, and 128 objects. In a rough approximation of the human data, the number of clusters was set to 2 log2 (n), and as in previous A DCLUS work, each object was inserted in each cluster with probability 0.5. A single similarity matrix was generated from each model using weights and constants uniformly distributed between 1 and 6. The third set of problems was derived from the second by adding gaussian noise with a variance of 10% of the variance of the similarity data and enforcing symmetry. Each algorithm was run at least 50 times on each data set. Runs that crashed or resulted in a VAF < 0 were ignored. To avoid biasing our conclusions in favor of methods requiring more computation time, those results were then used to derive the distribution of results that would be expected if all algorithms were run simultaneously and those that finished early were re-run repeatedly until the slowest algorithm finished its first run, with any re-runs in progress at that point discarded. 1 1 Depending as it does on running time, this comparison remains imprecise due to variations in the degree of code tuning and the quality of the compilers used, and the need to normalize timings between the multiple machines used in the tests. Summaries of the time-equated results produced by each algorithm on each of the human data sets are shown in Table 2. (em-indclus took much longer than the other algorithms and its performance is shown separately in Table 3.) The mean VAF for each algorithm is listed, along with the inter-quartile range (IQR) and the mean number of runs that were necessary to achieve time parity with the slowest algorithm on that data set (r). On most instances, there is remarkable variance in the VAF achieved by each algorithm. 2 Overall, despite the variety of approaches that have been brought to bear over the years, the original indclus algorithm appears to be the best. (Results in which another algorithm was superior to indclus are marked with a box.) Animals-s is the only data set on which its median performance was not the best, and its overall distribution of results is consistently competitive. It is revealing to note the differences in performance between the original and reimplemented versions of sindclus. Small changes in the handling of boundary cases make a large difference in the performance of the algorithm. Surprisingly, on the synthetic data sets (not shown), the relative performance of the algorithms was quite different, and almost the same on the noisy data as on the noise-free data. (This suggests that the randomly generated data sets that are commonly used to evaluate A DCLUS algorithms do not accurately reflect the problems of interest to practitioners.) ewindclus performed best here, although it was only occasionally able to recover the original models from the noise-free data. Overall, it appears that current methods of additive clustering are quite sensitive to the type of problem they are run on and that there is little assurance that they can recover the underlying structure in the data, even for small problems. To address these problems, we turn now to a new approach. 3 A Purely Combinatorial Approach One common theme in indclus, sindclus, and ewindclus is their computation of each cluster and its weight in turn, at each step fitting only the residual similarity not accounted for by the other clusters. This forces memberships to be considered in a predetermined order and allows weights to become obsolete. Inspired in part by recent work of Lee (in press), we propose an orthogonal decomposition of the problem. Instead of computing the elements and weight of each cluster in succession, we first consider all the memberships and then derive all the weights using constrained regression. And where previous algorithms recompute all the memberships of one cluster simultaneously (and therefore independently), we will change memberships one by one in a dynamically determined order using simple heuristic search techniques, recomputing the weights after each step. (An incremental bounded least squares regression algorithm that took advantage of the previous solution would be ideal, but the algorithms tested below did not incorporate such an improvement.) From this perspective, one need only focus on changing the binary membership variables, and A DCLUS becomes a purely combinatorial optimization problem. We will evaluate three different algorithms based on this approach, all of which attempt to improve a random initial model. The first, indclus-hc, is a simple hill-climbing strategy which attempts to toggle individual memberships in an arbitrary order and the first change resulting in an improved model is accepted. The algorithm terminates when no membership can be changed to give an improvement. This strategy is reminiscent of a proposal by Clouse and Cottrell (1996), although here we are using the A DCLUS model of similarity. In the second algorithm, indclus-pbil, the PBIL algorithm of Baluja (1997) is used 2 Table 3 shows one anomaly: no em-indclus run on animals resulted in a VAF ? 0. This also occurred on all synthetic problems with 32 or more objects (although very good solutions were found on the smaller problems). Tenenbaum (personal communication) suggests that the default annealing schedule in the em-indclus code may need to be modified for these problems. Table 4: The performance of the combinatorial algorithms on human data sets. indclus-hc Name animals-s numbers workers consonants animals letters VAF IQR 80 90 88 86 71 70 80?80 90?91 88?89 85?87 70?72 69?71 r 44 24 16 11 8 3 ind-pbil ewind-klb VAF IQR VAF IQR 74 87 86 80 66 66 71?74 85?88 84?87 76?82 65?69 64?68 80 91 89 92 74 76 80?80 91?91 89?89 92?92 74?74 74?78 indclus r 74 18 13 9 6 2 VAF IQR 80 90 88 91 74 82 80?80 89?91 88?89 91?91 74?74 81?82 r 47 59 53 61 36 57 to search for appropriate memberships. This is a simplification of the strategy suggested by Lee (in press), whose proposal also includes elements concerned with automatically controlling model complexity. We use the parameter settings he suggests but only allow the algorithm to generate 10,000 solutions. 3.1 KL Break-Out: A New Optimization Heuristic While the two approaches described above do not use any problem-specific information beyond solution quality, the third algorithm uses the gradient function from the ewindclus algorithm to guide the search. The move strategy is a novel combination of gradient ascent and the classic method of Kernighan and Lin (1970) which we call ?KL break-out?. It proceeds by gradient ascent, changing the entry in F whose ewindclus gradient points most strongly to the opposite of its current value. When the ascent no longer results in an improvement, a local maximum has been reached. Motivated by results suggesting that good maxima tend to cluster (Boese, Kahng, and Muddu, 1994; Ruml et al., 1996), the algorithm tries to break out of the current basin of attraction and find a nearby maximum rather than start from scratch at another random model. It selects the least damaging variable to change, using the gradient as in the ascent, but now it locks each variable after changing it. The pool of unlocked variables shrinks, thus forcing the algorithm out of the local maximum and into another part of the space. To determine if it has escaped, a new gradient ascent is attempted after each locking step. If the ascent surpasses the previous maximum, the current break-out attempt is abandoned and the ascent is pursued. If the break-out procedure changes all variables without any ascent finding a better maximum, the algorithm terminates. The procedure is guaranteed to return a solution at least as good as that found by the original KL method (although it will take longer), and it has more flexibility to follow the gradient function. This algorithm, which we will call ewindclus-klb, surpassed the original KL method in time-equated tests. It is also conceptually simple and has no parameters that need to be tuned. 3.2 Evaluation of the Combinatorial Algorithms The time-equated performance of the combinatorial algorithms on the human data sets is shown in Table 4, with indclus, the best of the previous algorithms, shown for comparison. As one might expect, adding heuristic guidance to the search helps it enormously: ewindclus-klb surpasses the other combinatorial algorithms on every problem. It performs better than indclus on three of the human data sets (top panel), equals its performance on two, and performs worse on one data set, letters. (Results in which ewindclus-klb was not the best are marked with a box.) The variance of indclus on letters is very small, and the full distributions suggest that ewindclus-klb is the better choice on this data set if one can afford the time to take the best of 20 runs. (Experiments Table 5: ewindclus-klb and indclus on noisy synthetic data sets of increasing size. ewindclus-klb indclus n VAF IQR VAF IQR r 8 97 96?97 95 93?97 1 16 91 90?92 86 85?87 4 32 90 88?92 83 82?84 22 64 91 90?91 84 84?85 100 128 91 91?91 88 87?90 381 using 7 additional human data sets found that letters represented the weakest performance of ewindclus-klb.) Performance of a plain KL strategy (not shown) surpassed or equaled indclus on all but two problems (consonants and letters), indicating that the combinatorial approach, in tandem with heuristic guidance, is powerful even without the new ?KL break-out? strategy. While we have already seen that synthetic data does not predict the relative performance of algorithms on human data very well, it does provide a test of how well they scale to larger problems. On noise-free synthetic data, ewindclus-klb reliably recovered the original model on all data sets. It was also the best performer on the noisy synthetic data (a comparison with indclus is presented in Table 5. These results show that, in addition to performing best on the human data, the combinatorial approach retains its effectiveness on larger problems. In addition to being able to handle larger problems than previous methods, it is important to note that the higher VAF of the models induced by ewindclus-klb often translates into increased interpretability. In the model shown in Table 1, for instance, the best previously published model (Tenenbaum, 1996), whose VAF is only 1.6% worse, does not contain s? in the unvoiced cluster. 4 Conclusions We formalized the problem of constructing feature-based representations for cognitive modeling as the unsupervised learning of A DCLUS models from similarity data. In an empirical comparison sensitive to variance in solution quality and computation time, we found that several recently proposed methods for recovering such models perform worse than the original indclus algorithm of Arabie and Carroll (1980). We suggested a purely combinatorial approach to this problem that is simpler than previous proposals, yet more effective. By changing memberships one at a time, it makes fewer independence assumptions. We also proposed a novel variant of the Kernighan-Lin optimization strategy that is able to follow the gradient function more closely, surpassing the performance of the original. While this work has extended the reach of the additive clustering paradigm to large problems, it is directly applicable to feature construction of only those cognitive models whose representations encode similarity as shared features. (The cluster weights can be represented by duplicating strong features or by varying connection weights.) However, the simplicity of the combinatorial approach should make it straightforward to extend to models in which the absence of features can enhance similarity. Other future directions include using the output of one algorithm as the starting point for another, and incorporating measures of model complexity(Lee, in press). 5 Acknowledgments Thanks to Josh Tenenbaum, Michael Lee, and the Harvard AI Group for stimulating discussions; to Josh, Anil Chaturvedi, Henk Kiers, J. Douglas Carroll, and Phipps Arabie for providing source code for their algorithms; Josh, Michael, and Phipps for providing data sets; and Michael for sharing unpublished work. This work was supported in part by the NSF under grants CDA-94-01024 and IRI-9618848. References Arabie, Phipps and J. Douglas Carroll. 1980. MAPCLUS: A mathematical programming approach to fitting the adclus model. Psychometrika, 45(2):211?235, June. Baluja, Shumeet. 1997. Genetic algorithms and explicit search statistics. In Michael C. Mozer, Michael I. Jordan, and Thomas Petsche, editors, NIPS 9. Boese, Kenneth D., Andrew B. Kahng, and Sudhakar Muddu. 1994. A new adaptive multi-start technique for combinatorial global optimizations. Operations Research Letters, 16:101?113. Carroll, J. Douglas and Phipps Arabie. 1983. INDCLUS: An individual differences generalization of the ADCLUS model and the MAPCLUS algorithm. Psychometrika, 48(2):157?169, June. Chaturvedi, Anil and J. Douglas Carroll. 1994. An alternating combinatorial optimization approach to fitting the INDCLUS and generalized INDCLUS models. Journal of Classification, 11:155?170. Clouse, Daniel S. and Garrison W. Cottrell. 1996. Discrete multi-dimensional scaling. In Proceedings of the 18th Annual Conference of the Cognitive Science Society, pp. 290?294. Hojo, Hiroshi. 1983. A maximum likelihood method for additive clustering and its applications. Japanese Psychological Research, 25(4):191?201. Kernighan, B. and S. Lin. 1970. An efficient heuristic procedure for partitioning graphs. The Bell System Technical Journal, 49(2):291?307, February. Kiers, Henk A. L. 1997. A modification of the SINDCLUS algorithm for fitting the ADCLUS and INDCLUS models. Journal of Classification, 14(2):297?310. Lee, Michael D. in press. A simple method for generating additive clustering models with limited complexity. Machine Learning. Mechelen, I. Van and G. Storms. 1995. Analysis of similarity data and Tversky?s contrast model. Psychologica Belgica, 35(2?3):85?102. Noelle, David C., Garrison W. Cottrell, and Fred R. Wilms. 1997. Extreme attraction: On the discrete representation preference of attractor networks. In M. G. Shafto and P. Langley, eds, Proceedings of the 19th Annual Conference of the Cognitive Science Society, p. 1000. Ruml, Wheeler, J. Thomas Ngo, Joe Marks, and Stuart Shieber. 1996. Easily searched encodings for number partitioning. Journal of Optimization Theory and Applications, 89(2). Shepard, Roger N. and Phipps Arabie. 1979. Additive clustering: Representation of similarities as combinations of discrete overlapping properties. Psychological Review, 86(2):87?123, March. Stark, Philip B. and Robert L. Parker. 1995. Bounded-variable least-squares: An algorithm and applications. Computational Statistics, 10:129?141. Tenenbaum, Joshua B. 1996. Learning the structure of similarity. In D. S. Touretzky, M. C. Mozer, and M. E. Hasselmo, editors, NIPS 8.
2077 |@word version:4 tedious:1 pbil:3 decomposition:1 initial:2 configuration:2 contains:1 selecting:1 daniel:1 tuned:1 genetic:1 existing:2 current:4 recovered:1 activation:1 assigning:1 yet:1 must:1 reminiscent:1 cottrell:5 additive:16 predetermined:1 pursued:1 fewer:2 leaf:1 assurance:1 obsolete:1 mental:1 recompute:1 contribute:1 preference:1 simpler:2 mathematical:2 along:2 become:2 ik:1 fitting:5 pairwise:2 inter:1 expected:1 behavior:1 gjk:1 multi:2 inspired:1 automatically:3 little:1 considering:1 increasing:1 becomes:1 tandem:1 psychometrika:2 underlying:1 formalizing:1 bounded:2 panel:1 unspecified:1 developed:1 finding:2 guarantee:1 duplicating:1 every:1 vaf:22 tie:1 partitioning:2 unit:2 grant:1 engineering:1 shumeet:1 timing:1 local:2 phipps:5 limit:1 despite:2 encoding:1 oxford:1 might:1 quantified:1 dynamically:1 specifying:1 suggests:3 limited:1 range:1 practical:1 acknowledgment:1 practice:1 union:1 reimplementation:1 differs:1 procedure:3 wheeler:2 langley:1 empirical:4 bell:1 revealing:1 imprecise:1 suggest:2 optimize:1 equivalent:1 straightforward:1 iri:1 starting:2 independently:1 simplicity:2 formalized:1 fik:2 attraction:2 classic:1 handle:3 variation:1 mapclus:2 controlling:1 construction:1 anomaly:1 programming:2 us:1 harvard:3 element:3 asymmetric:3 cloud:1 inserted:1 highest:1 principled:1 mozer:2 predictable:1 complexity:3 locking:1 arabie:11 personal:1 tversky:1 solving:1 purely:3 division:1 easily:1 represented:2 fast:2 effective:2 hiroshi:1 artificial:1 whose:5 heuristic:8 larger:4 valued:3 encoded:1 solve:1 widely:1 otherwise:1 quite:2 objectively:1 favor:1 statistic:2 noisy:3 ptk:1 higherlevel:1 advantage:1 took:2 propose:2 chaturvedi:4 flexibility:1 achieve:1 normalize:1 cluster:23 empty:1 produce:2 generating:1 incremental:1 object:17 help:1 derive:2 depending:1 andrew:1 ij:3 progress:1 strong:1 recovering:1 involves:1 direction:1 posit:1 guided:1 closely:1 shafto:1 quartile:1 human:10 require:1 generalization:1 articulate:1 considered:1 predict:1 early:3 klb:11 applicable:1 combinatorial:15 sensitive:2 hasselmo:1 noelle:2 rough:1 clearly:1 brought:1 always:2 gaussian:1 modified:4 rather:1 avoid:1 fricative:1 varying:2 encode:1 derived:4 focus:1 june:2 improvement:3 consistently:1 likelihood:1 slowest:2 contrast:1 equaled:1 sense:1 membership:15 typically:1 diminishing:1 hidden:1 comprising:1 selects:1 overall:3 classification:2 animal:10 constrained:2 equal:1 construct:3 having:1 sampling:1 identical:1 stuart:1 unsupervised:2 future:1 connectionist:2 report:2 randomly:1 dg:1 simultaneously:2 resulted:2 individual:3 intended:2 connects:1 attractor:1 attempt:3 interest:1 investigate:2 evaluation:2 extreme:1 yielding:1 worker:3 necessary:1 orthogonal:1 desired:1 re:3 guidance:2 cda:1 psychological:4 increased:2 formalism:2 modeling:3 column:1 instance:5 recomputing:1 retains:1 assignment:2 maximization:1 phrase:1 surpasses:2 subset:2 entry:1 front:2 reported:3 eec:1 synthetic:8 notate:1 thanks:1 probabilistic:1 lee:5 pool:1 enhance:1 michael:6 reflect:1 containing:1 worse:3 cognitive:6 return:2 stark:2 account:1 converted:1 suggesting:1 wk:4 includes:1 later:1 try:2 performed:1 break:6 compiler:1 competitive:1 recover:2 maintains:1 reached:1 start:2 substructure:1 voiced:3 square:3 accuracy:1 variance:8 succession:1 climbing:1 conceptually:1 accurately:1 produced:1 published:2 reach:1 touretzky:1 manual:1 sharing:1 ed:1 pp:1 storm:2 stop:3 formalize:1 schedule:1 back:1 appears:2 originally:1 higher:1 follow:2 improved:1 formulation:3 box:2 strongly:1 shrink:1 furthermore:1 just:1 stage:3 roger:1 until:1 hand:3 expressive:1 overlapping:2 lack:1 kernighan:3 quality:6 perhaps:1 name:3 effect:1 requiring:1 contain:1 alternating:3 semantic:1 ind:1 maintained:1 criterion:1 generalized:1 toggle:1 hill:1 performs:3 wise:1 novel:4 recently:1 superior:1 common:1 shepard:4 extend:1 interpretation:1 occurred:1 he:1 surpassing:1 significant:1 cambridge:1 gibbs:1 ai:1 automatic:1 tuning:1 language:1 similarity:18 carroll:9 longer:3 showed:1 recent:1 perspective:1 forcing:1 manipulation:1 occasionally:1 binary:3 arbitrarily:1 joshua:1 seen:1 additional:2 performer:1 converge:1 determine:1 paradigm:1 afterwards:1 multiple:3 full:1 technical:1 match:1 iqr:13 lin:3 escaped:1 coded:1 ensuring:1 variant:1 regression:4 noiseless:1 expectation:1 surpassed:2 iteration:1 represent:4 achieved:2 proposal:3 addition:2 separately:1 annealing:2 median:1 source:1 eliminates:1 anp:1 ascent:8 induced:1 tend:1 effectiveness:2 jordan:1 practitioner:1 call:2 ngo:1 ideal:1 constraining:1 kiers:3 concerned:1 variety:1 independence:3 fit:1 psychology:1 opposite:1 translates:1 motivated:1 effort:1 penalty:1 reformulated:1 cause:1 afford:1 repeatedly:2 ignored:1 clear:1 listed:2 nasal:1 factorial:1 amount:1 tenenbaum:5 generate:1 nsf:1 discrete:7 promise:1 group:1 changing:4 douglas:4 clean:1 kenneth:1 graph:1 merely:1 sum:1 year:1 run:15 letter:8 powerful:1 almost:1 scaling:2 guaranteed:1 simplification:1 quadratic:1 annual:2 precisely:1 nearby:1 performing:1 wilms:2 combination:3 march:1 across:1 slightly:1 em:8 terminates:3 smaller:1 appealing:1 making:1 modification:2 intuitively:1 gradually:1 sij:2 taken:2 previously:2 remains:1 abbreviated:1 turn:5 operation:1 hierarchical:2 appropriate:1 petsche:1 original:13 abandoned:1 assumes:1 clustering:16 running:3 top:1 include:1 thomas:2 log2:1 lock:1 february:1 society:2 move:1 arrangement:2 realized:1 already:1 strategy:7 unclear:1 gradient:10 simulated:1 street:1 philip:1 considers:1 toward:1 enforcing:1 code:5 modeled:1 relationship:1 providing:3 difficult:3 unfortunately:1 robert:1 negative:2 synthesizing:1 design:2 implementation:4 kahng:2 reliably:1 perform:1 unvoiced:5 discarded:1 descent:1 extended:1 communication:1 arbitrary:1 david:1 complement:1 crashed:1 unpublished:1 kl:6 extensive:1 optimized:2 connection:1 adclus:3 established:1 nip:2 address:2 beyond:2 able:3 suggested:2 below:7 pattern:1 reimplemented:2 proceeds:1 biasing:1 interpretability:1 confusability:1 difficulty:1 treated:1 force:1 predicting:1 abbreviate:1 residual:1 mn:1 improve:1 shieber:1 finished:2 superseded:1 fjk:1 review:2 determining:1 relative:2 fully:1 encompassing:1 bear:1 expect:1 clouse:3 remarkable:1 degree:2 basin:1 principle:1 editor:2 share:2 summary:1 accounted:2 surprisingly:1 parity:1 free:3 changed:1 supported:1 guide:1 allow:1 taking:1 distributed:2 van:1 boundary:2 default:1 plain:1 fred:1 equated:3 collection:1 commonly:1 adaptive:1 simplified:1 global:1 consonant:5 search:6 table:13 symmetry:1 hc:2 japanese:1 constructing:5 domain:1 ruml:4 did:2 noise:4 henk:2 crafted:1 enormously:1 parker:2 garrison:2 theme:1 explicit:2 psychologica:1 third:2 anil:2 specific:2 concern:1 weakest:1 burden:1 incorporating:1 joe:1 adding:2 effectively:1 intersection:1 simply:2 josh:3 ma:1 stimulating:1 marked:2 shared:2 absence:2 change:7 determined:2 typical:2 except:1 uniformly:1 justify:1 wt:1 baluja:2 pas:1 accepted:1 attempted:1 indicating:1 damaging:1 mark:1 searched:1 incorporate:1 evaluate:2 tested:1 scratch:1 handling:2
1,182
2,078
A Generalization of Principal Component Analysis to the Exponential Family Michael Collins  Sanjoy Dasgupta Robert E. Schapire AT&T Labs Research 180 Park Avenue, Florham Park, NJ 07932 mcollins, dasgupta, schapire  @research.att.com Abstract Principal component analysis (PCA) is a commonly applied technique for dimensionality reduction. PCA implicitly minimizes a squared loss function, which may be inappropriate for data that is not real-valued, such as binary-valued data. This paper draws on ideas from the Exponential family, Generalized linear models, and Bregman distances, to give a generalization of PCA to loss functions that we argue are better suited to other data types. We describe algorithms for minimizing the loss functions, and give examples on simulated data. 1 Introduction Principal component analysis (PCA) is a hugely popular dimensionality reduction technique that attempts to find a low-dimensional subspace passing close to a given set of points        . More specifically, in PCA, we find a lower dimensional subspace that  minimizes the sum of the squared distances from the data points  to their projections  in the subspace, i.e.,              (1) This turns out to be equivalent to choosing a subspace that maximizes the sum of the squared lengths of the projections  , which is the same as the (empirical) variance of these projections if the data happens to be centered at the origin (so that     ). PCA also has another convenient interpretation that is perhaps less well known. In this probabilistic interpretation, each point  is thought of as a random draw from some un known distribution "! , where  denotes a unit Gaussian with mean #$ . The purpose   then of PCA is to find the set of parameters %     that maximizes the likelihood of the data, subject to the condition that these parameters all lie in a low-dimensional subspace. In         are considered to be noise-corrupted versions of some true points other words,       & which lie in a subspace; the goal is to find these true points, and the main assumption is that the noise is Gaussian. The equivalence of this interpretation to the ones given above follows simply from the fact that negative log likelihood under this Guassian model is equal (ignoring constants) to Eq. (1). This Gaussian assumption may be inappropriate, for instance if data is binary-valued, or integer-valued, or is nonnegative. In fact, the Gaussian is only one of the canonical distributions that make up the exponential family, and it is a distribution tailored to real-valued data. The Poisson is better suited to integer data, and the Bernoulli to binary data. It seems natural to consider variants of PCA which are founded upon these other distributions in place of the Gaussian.    be any parameterized set We extend PCA to the rest of the exponential family.  Let of distributions from the exponential family, where is the natural parameter of a distribution. For instance, a one-dimensional Poisson distribution can be parameterized by  ,   and distribution corresponding to mean               Given data      &    , the goal is now to find parameters        !   which lie in a low-dimensional subspace and for which the log-likelihood   is maximized.                      Our unified approach effortlessly permits hybrid dimensionality reduction schemes in which different types of distributions can be used for different attributes of the data. If the data   have a few binary attributes and a few integer-valued attributes, then some co ordinates of the corresponding  can be parameters of binomial distributions while others are parameters of Poisson distributions. (However, for simplicity of presentation, in this abstract we assume all distributions are of the same type.) The dimensionality reduction schemes for non-Gaussian distributions are substantially different from PCA. For instance, in PCA the parameters  , which are means of Gaussians, lie in a space which coincides with that of the data   . This is not the case in general, and therefore, although the parameters  lie in a linear subspace, they typically correspond to a nonlinear surface in the space of the data.  The discrepancy and interaction between the space of parameters and the space of the data  is a central preoccupation in the study of exponential families, generalized linear models (GLM?s), and Bregman distances. Our exposition is inevitably woven around these three intimately related subjects. In particular, we show that the way in which we generalize PCA is exactly analogous to the manner in which regression is generalized by GLM?s. In this respect, and in others which will be elucidated later, it differs from other variants of PCA recently proposed by Lee and Seung [7], and by Hofmann [4]. We show that the optimization problem we derive can be solved quite naturally by an algorithm that alternately minimizes over the components of the analysis and their coefficients; thus, the algorithm is reminiscent of Csisz?ar and Tusn?ady?s alternating minization procedures [2]. In our case, each side of the minimization is a simple convex program that can be interpreted as a projection with respect to a suitable Bregman distance; however, the overall program is not generally convex. In the case of Gaussian distributions, our algorithm coincides exactly with the power method for computing eigenvectors; in this sense it is a generalization of one of the oldest algorithms for PCA. Although omitted for lack of space, we can show that our procedure converges in that any limit point of the computed coefficients is a stationary point of the loss function. Moreover, a slight modification of the optimization criterion guarantees the existence of at least one limit point. ! #" $ &% Some comments on notation: All vectors in this paper are row vectors. If we denote its ?th row by  and its ?th element by  .  is a matrix, 2 The Exponential Family, GLM?s, and Bregman Distances 2.1 The Exponential Family and Generalized Linear Models In the exponential family of distributions the conditional probability of a value parameter value takes the following form:   '   ( ) +*  ' -, '. 0/   ' given (2) ' /   / '  ()  * '   We use to denote the domain of ' . The  sum is replaced by an integral in the continuous case, where defines a density over . * is a term that depends only on ' , and can usually be ignored as a constant during estimation. difference between different  . We willTheseemain that almost all of the concepts of members of the family is the form of / the PCA algorithms in this paper stem directly from the definition of / . A first example is a normal distribution,  and unit variance, which has a density  '    with()mean that that is usually written as       '     . It can be verified     this is  a member of the exponential family with , * '  '  ,               . Another and / common case is a Bernoulli distribution for the case of binary       . The probability of '   is usually written  '  +  case outcomes.   -  In thiswhere    . This is a member of the exponential family is a parameter in    ,      . with * '   ,  ()  , and /  /  , which we will denote as   throughout this A critical function is the derivative  /      * '   , it is easily verified  paper.  that     '   ,Bythedifferentiating  expectation of under . In the normal distribution, , and in  '   ' '   . In the general case,   '   is referred to as the ?expectation the Bernoulli case  '   parameter?, and  defines a function from the natural parameter values to the expectation Here, is the ?natural parameter? of the distribution, and can usually take any value in the reals. is a function that ensures that the sum (integral) of over the domain of   is 1. From this it follows that . parameter values. Our generalization of PCA is analogous to the manner in which generalized linear models (GLM?s) [8] provide a unified treatment of regression for the exponential family by generalizing least-squares regression to loss functions that are more appropriate for other members of this family. The regression set-up assumes a training sample of     pairs, where      is a vector of attributes, and    is some response variable. The pa   . The dot product  is taken to be an rameters of the model are a vector are set to be approximation of  . In least squares regression the optimal parameters  &  .               !" $#&%('   *)  +  In GLM?s, , -  is taken to approximate the expectation parameter of the exponential model, where , is the inverse of the ?link function? [8].  A natural choice is to use the ?canonical link?, where ,   ,  being the derivative /. . In this case the natural   /     parameters are directly approximated by , and the log-likelihood     '   ,  / 2   . In the case of a normal distribution is simply   () *     01    /    and it follows easily that the maximum-likelihood criwith fixed variance, terion is equivalent to the least squares Another interesting case is logistic re  , , andcriterion.     gression3 where / the negative log-likelihood for parameters  is      ! 7 8 !:9 , where      if     ,      if     .     ) 546 2.2 Bregman Distances and the Exponential Family ;2<>=@?  be a differentiable and strictly convex function defined on a closed, convex =BA  . The Bregman distance associated with ; is defined for  C D= to be EGF H JI C   ; H + ;  C LK  C M C   where K '  ;  ' . It can be shown that, in general, every Bregman distance is nonnegative and is equal to zero if and only if its two arguments are equal.  '  is directly related to a Bregman For the exponential family the log-likelihood () Let set normal   $   , . -/ 56.-/87 $39 ' .-3 <>=  ?A@"B0 -   < = I-A@ .- $ 0J Poisson      !# ! # .-/ 0 1   ? B ? 0GD FE 1 -N ! # 1 0 1  - C? Bernoulli       "!# % &('*) %  -3  -  / .+ -/ 0 21 0 41  ;' 9 : ' ?  ?/ : 0ED FG H1 0 ' 99 D F   -LKH7    "! 9 : # where M 1  B0   1 0   1 6 Table 1: Various functions of interest for three members of the exponential family / and  : ;   , /    ;  through (3)    K It can be shown under fairly general conditions that '   ' . Application of these identities implies that the negative log-likelihood of a point can be expressed through a Bregman distance [1, 3]: (4) ()  '   () +*  ' ;  ' , E F  ' I    distance. Specifically, [1, 3] define a ?dual? function In other words, negative log-likelihood can always be written as a Bregman distance plus a term that is constant with respect to and which therefore can be ignored. Table 1 summarizes various functions of interest for examples of the exponential family. We will find it useful to extend the idea of Bregman distances to divergences between vectheT notation tors and matrices. If  , O are vectors, and P , Q are matrices, then we overload SR U      and       . (The as  O P Q notion of Bregman distance as well as our generalization of PCA can be extended to vectors in a more general manner; here, for simplicity, we restrict our attention to Bregman distances and PCA problems of this particular form.) EF  I E F ' I  EF  I %E F  % I % 3 PCA for the Exponential Family  We now generalize PCA to other members of the exponential family. We wish to find  ?s that are ?close? to the  ?s and which belong to a lower dimensional subspace of parameter  space. Thus, our approach is to find a basis  V %   R   Y VXW Y in  and to represent each  as the linear combination of these elements   ZY  V that is ?closest? to   . ]\_^   \b^ Let [ be the matrix matrix whose d\ whose ?th row is   . LetR ` be the a d\fc ^ ?th row is V Y , and let P be the matrix a matrix with elements  Y . Then e  PG` is an whose ?th row is  as above. This is a matrix of natural parameter values which define the probability of each point in [ . `   [  ' &%  &%      % ,  ' &%  % , /   % % Following the discussion in Section 2, we consider the loss function taking the form g P   P `     Zh    where h is a constant term which will be dropped from here on. The loss function varies depending on which member of the exponential family is taken, which simply changes the form of . For example, if [ is a matrix  of real values, and the normal distribution is  appropriate for the data, then and the loss criterion is the usual squared loss   for PCA. For the Bernoulli distribution, . If we define   ,  ! i  ! i g    then `  P .  / /     , /   )  , %-   6  ' & %  ' &%  From the relationship between log-likelihood and Bregman distances (see Eq. (4)), the loss can also be written as g ` P      E F  ' &% I    % %    E F    I      (where we allow to be applied to vectors and matrices in a pointwise manner). Once ` and P have been found for the data points, the ?th data point    can be represented W as the vector  in the lower dimensional space  . Then  are the coefficients which define a Bregman projection of the vector   :   "   &#  % '  E F    I   3`  (5) The generalized form of PCA can also be considered to be search for a low dimensional basis (matrix ` ) which defines a surface that is close to all the data points   . We define the      3`   W . The optimal value for ` then minset of points ` to be `  imizes the sum of projection distances: `   .       and the Bregman distance is Euclidean disNote that for the normal distribution tance so that the projection operation in Eq. (5) is a simple linear projection ("    ` ).  is also simplified in the normal case, simply being the hyperplane whose basis is ` . `       / #&%('   E F  I "  # %(' To summarize, once a member of the exponential family ? and by implication a convex function ? is chosen, regular PCA is generalized in the following way:   The loss function is negative log-likelihood,  constant.     ()  ' ' , /  -, The matrix e  PG` is taken to be a matrix of natural parameter values.  of /  defines a matrix of expectation parameters,   PG` . The derivative  E F is derived from ; . A function ; is derived from / and  . A Bregman distance  The loss is a sum of Bregman distances from the elements '  % to values  % % .    PCA can also be thought of as search for a matrix ` is ?close? to all the data points.  ` that defines a surface which   , and The normal distribution is a simple case because  the divergence is Euclidean distance. The projection operation is a linear operation, and ` is the hyperplane which has ` as its basis. 4 Generic Algorithms for Minimizing the Loss Function  We now describe a generic algorithm for minimization of the loss function. First, we concentrate on the simplest case where there is just a single component so that a  . (We drop R the c subscript from  Y and  Y .) The method is iterative, with an initial random choice for   the value of ` . Let ` ,P , etc. denote the values at the  ?th  iteration, and let g `   be the  P `  P initial random choice. We propose the iterative updates     g g  and ` . Thus is alternately minimized with respective to `  P  its two arguments, each time optimizing one argument while keeping the other one fixed, reminiscent of Csisz?ar and Tusn?ady?s alternating minization procedures [2].   % "  # %('    "  # %(' 3 9 % E E F F 3 ' &% I I  S*R R   %   9  ' %    . It is useful to write these minimization problems as follows: R  For     ,       ^  For     ,        "   !"  ! " $#&%('  % $ #&%('  *    , ^ We can then see that there are optimization problems, and that each one is essentially identical to a GLM regression problem (a very simple one, where there is a single parameter being optimized over). These sub-problems are easily solved, as the functions are convex in the argument being optimized over, and the large literature on maximumlikelihood estimation in GLM?s can be directly applied to the problem.            These updates take a simple form for the nor                   P mal distribution: , and ` . It ` [ P  P    [ `  ` [  [ h , where h is a scalar value. The method is then equivfollows that ` alent to the power method (see Jolliffe [5]) for finding the eigenvector of [  [ with the largest eigenvalue, which is the best single component solution for ` . Thus the generic algorithm generalizes one of the oldest algorithms for solving the regular PCA problem.   The loss is convex in either of its arguments with the other fixed, but in general is not convex in the two arguments together. This makes it very difficult to prove convergence to the global minimum. The normal distribution is an interesting special case in this respect ? the power method is known to converge to the optimal solution, in spite of the non-convex nature of the loss surface. A simple proof of this comes from properties of eigenvectors (Jolliffe [5]). It can also be explained by analysis of the Hessian : for any stationary point is not positive semi-definite. Thus these stationary which is not the global minimum, points are saddle points rather than local minima. The Hessian for the generalized loss function is more complex; it remains an open problem whether it is also not positive semidefinite at stationary points other than the global minimum. It is also open to determine under which conditions this generic algorithm will converge to a global minimum. In preliminary numerical studies, the algorithm well behaved in this respect.  seems to be   P ` Moreover, any limit point of the sequence e will be a stationary point.      However, it is possible for this sequence to diverge since the optimum may be at infinity. To avoid such degenerate choices of e , we can use a modified loss E F  ' &% I    % , E F   * I    %    %  where is a small   positive constant, and +* is any value in the range of  (and therefore for which   * is finite). This is roughly equivalent to adding a conjugate prior and finding the maximum solution. It can be proved, for this modified loss, that the   remainsa posteriori sequence e in a bounded region and hence always has at least one limit point   which must be a stationary point. (All proofs omitted for lack of space.) There are various ways to optimize the loss function when there is more than one component. We give one algorithm which cycles through the a components, optimizing each in turn while the others are held fixed:         ! / 10 //Initialization 7 7 Set , //Cycle through components times 7 7 0    , 0    : For //Now optimize & the ?th component with other components fixed 7 Initialize + randomly, and set  7 For 0    convergence & & 9 ' 7 7 @ $       ,  For  < = + +  & 7 7 @ $  & < = 8   ,  For + +             3 9  ##"" $% '&)(#*,+-  3     "  #" . 9      $% '&)(#*32 -        .    The modified Bregman projections now include a term 4 &% representing the contribution of the a  fixed components. These sub-problems are again a standard optimization problem regarding Bregman distances, where the terms 4  % form a ?reference prior?. 90 500 data PCA exp 80 data PCA exp 450 70 400 60 350 300 50 Y Y 40 250 200 30 150 20 100 10 50 0 0 10 20 30 40 50 60 70 80 0 0 90 20 40 60 X 80 100 X Figure 1: Regular PCA vs. PCA for the exponential distribution. B B 1 1 0.8 0.8 B? 0.6 0.4 0.4 0.2 0.2 0 1 0 1 A B? 0.6 C? C 0.8 A 1 0.6 D C 0.8 1 0.6 0.8 0.6 0.4 0.4 0.2 0.2 0 0.8 0.6 0.4 0.4 0.2 C? D? 0.2 0 0 0 E Figure 2: Projecting from 3- to 1-dimensional space, via Bernoulli PCA. Left: the three points   h are projected onto a one-dimensional curve. Right: point  is added. 5 Illustrative examples Exponential distribution. Our generalization of PCA behaves rather differently for different members of the exponential family. One interesting example is that of the exponential distributions on nonnegative reals. For one-dimensional data, these densities are usually  written as   , where is the mean. In the uniform system of notation we have been  using, we would instead index each distribution by a single natural parameter      , where  (basically,  .  ), and write the density  as  The link function in this case is , the mean of the distribution.      '     /   '   and want to find the best one-dimensional apSuppose we are given data [   *R V   proximation: a vector V and coefficients such that the approximation      &  has minimum loss. The alternating minimization procedure of the previous section has a simple closed form in this case, consisting of the iterative update rule    V   ^ [  [b V     Here the shorthand  denotes a componentwise reciprocal, i.e.,        . Notice the [  [bV . Once V is found, similarity to the update rule of the power for PCA: V ^ method R     The points    V lie on a line through we can recover the coefficients   toSR alsoR lie on a straight line; the origin. Normally, we would not expect the points however, in this case they do, because any point of the form V    , can be written    as   and so must lie in the direction  .      ) Therefore, we can reasonably ask how the lines found under this exponential assumption differ from those found under a Gaussian assumption (that is, those found by regular PCA), provided all data is nonnegative. As a  very simple illustration, we conducted two toy experiments with twenty data points in  (Figure 1). In the first, the points all lay very close to a line, and the two versions of PCA produced similar results. In the second experiment, a few of the points were moved farther afield, and these outliers had a larger effect upon regular PCA than upon its exponential variant. Bernoulli distribution. For the Bernoulli distribution, a linear subspace of the space of parameters is typically a nonlinear surface in the  space of the data. In Figure 2 (left), three points in the three-dimensional hypercube   are mapped via our PCA to a onedimensional curve. The curve passes through one of the points ( ); the projections of the h ) are indicated. Notice that the curve is symmetric about two other ( and h the center of the hypercube, . In Figure 2 (right), another point (D) is added,   and causes the approximating one-dimensional curve to swerve closer to it. E ? E   ?           6 Relationship to Previous Work Lee and Seung [6, 7] and Hofmann [4] also describe probabilistic alternatives to PCA, tailored to data types that are not gaussian. In contrast to our method, [4, 6, 7] approximate mean parameters underlying the generation of the data points, with constraints on the matrices P and ` ensuring that the elements of PG` are in the correct domain. By instead choosing to approximate the natural parameters, in our method the matrices P and ` do not usually need to be constrained?instead, we rely on the link function to give a transformed matrix PG` which lies in the domain of the data points.   %  ' &% () % , &%    More specifically, Lee and Seung [6] use the loss function    R (ignoring constant factors, and again defining  $  Y  Y  Y ). This is optimized with the constraint that P and ` should be positive. This method has a probabilistic interpretation, where each data point  is generated from a Poisson distribution with mean  parameter !i  .   For the Poisson distribution, our method uses the loss function    , but without any constraints on the matrices P and ` . The algorithm in Hofmann [4] uses   , where the matrices P and ` are constrained such that a loss function    all the  ?s are positive, and also such that     . ' &% % % ' &% () % &% % % % ' % &% ,   % %  Bishop and Tipping [9] describe probabilistic variants of the gaussian case. Tipping [10] discusses a model that is very similar to our case for the Bernoulli family. Acknowledgements. This work builds upon intuitions about exponential families and Bregman distances obtained largely from interactions with Manfred Warmuth, and from his papers. Thanks also to Andreas Buja for several helpful comments. References [1] Katy S. Azoury and M. K. Warmuth. Relative loss bounds for on-line density estimation with the exponential family of distributions. Machine Learning, 43:211?246, 2001. [2] I. Csisz?ar and G. Tusn?ady. Information geometry and alternating minimization procedures. Statistics and Decisions, Supplement Issue, 1:205?237, 1984. [3] J?urgen Forster and Manfred Warmuth. Relative expected instantaneous loss bounds. Journal of Computer and System Sciences, to appear. [4] Thomas Hofmann. Probabilistic latent semantic indexing. In Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, 1999. [5] I. T. Jolliffe. Principal Component Analysis. Springer-Verlag, 1986. [6] D. D. Lee and H. S. Seung. Learning the parts of objects with nonnegative matrix factorization. Nature, 401:788, 1999. [7] Daniel D. Lee and H. Sebastian Seung. Algorithms for non-negative matrix factorization. In Advances in Neural Information Processing Systems 13, 2001. [8] P. McCullagh and J. A. Nelder. Generalized Linear Models. CRC Press, 2nd edition, 1990. [9] M. E. Tipping and C. M. Bishop. Probabilistic principal component analysis. Journal of the Royal Statistical Society, Series B, 61(3):611?622, 1999. [10] Michael E. Tipping. Probabilistic visualisation of high-dimensional binary data. In Advances in Neural Information Processing Systems 11, pages 592?598, 1999.
2078 |@word version:2 seems:2 nd:2 open:2 pg:5 reduction:4 initial:2 series:1 att:1 daniel:1 terion:1 com:1 reminiscent:2 written:6 must:2 numerical:1 hofmann:4 afield:1 drop:1 update:4 v:1 stationary:6 warmuth:3 oldest:2 reciprocal:1 farther:1 manfred:2 vxw:1 prove:1 shorthand:1 manner:4 expected:1 roughly:1 nor:1 inappropriate:2 provided:1 moreover:2 notation:3 maximizes:2 bounded:1 underlying:1 interpreted:1 minimizes:3 substantially:1 eigenvector:1 unified:2 finding:2 nj:1 guarantee:1 every:1 alent:1 exactly:2 unit:2 normally:1 appear:1 positive:5 dropped:1 local:1 limit:4 subscript:1 plus:1 initialization:1 equivalence:1 co:1 factorization:2 range:1 definite:1 differs:1 procedure:5 empirical:1 thought:2 projection:11 convenient:1 word:2 regular:5 spite:1 onto:1 close:5 optimize:2 equivalent:3 center:1 attention:1 hugely:1 convex:9 sigir:1 simplicity:2 rule:2 his:1 imizes:1 notion:1 analogous:2 mcollins:1 us:2 origin:2 pa:1 element:5 approximated:1 lay:1 solved:2 mal:1 region:1 ensures:1 cycle:2 intuition:1 seung:5 woven:1 solving:1 upon:4 basis:4 easily:3 differently:1 various:3 represented:1 guassian:1 describe:4 choosing:2 outcome:1 quite:1 whose:4 larger:1 valued:6 florham:1 statistic:1 sequence:3 differentiable:1 eigenvalue:1 propose:1 interaction:2 product:1 degenerate:1 moved:1 csisz:3 convergence:2 optimum:1 converges:1 object:1 derive:1 depending:1 b0:2 eq:3 implies:1 come:1 differ:1 concentrate:1 direction:1 correct:1 attribute:4 centered:1 crc:1 generalization:6 preliminary:1 strictly:1 effortlessly:1 around:1 considered:2 normal:9 exp:2 tor:1 omitted:2 purpose:1 estimation:3 largest:1 minimization:5 gaussian:10 always:2 modified:3 rather:2 avoid:1 derived:2 bernoulli:9 likelihood:11 contrast:1 sense:1 posteriori:1 helpful:1 letr:1 typically:2 visualisation:1 transformed:1 overall:1 dual:1 issue:1 development:1 constrained:2 special:1 fairly:1 initialize:1 urgen:1 equal:3 ady:3 once:3 identical:1 park:2 discrepancy:1 minimized:1 others:3 few:3 randomly:1 divergence:2 replaced:1 geometry:1 consisting:1 attempt:1 interest:2 semidefinite:1 held:1 implication:1 bregman:21 integral:2 closer:1 respective:1 euclidean:2 re:1 instance:3 ar:3 uniform:1 conducted:1 varies:1 corrupted:1 gd:1 thanks:1 density:5 international:1 probabilistic:7 lee:5 diverge:1 michael:2 together:1 squared:4 central:1 again:2 derivative:3 toy:1 coefficient:5 depends:1 later:1 h1:1 lab:1 closed:2 recover:1 contribution:1 square:3 variance:3 largely:1 maximized:1 correspond:1 generalize:2 zy:1 basically:1 produced:1 straight:1 sebastian:1 ed:1 definition:1 naturally:1 associated:1 proof:2 proved:1 treatment:1 popular:1 ask:1 dimensionality:4 tipping:4 response:1 just:1 nonlinear:2 lack:2 defines:5 logistic:1 indicated:1 perhaps:1 behaved:1 effect:1 concept:1 true:2 hence:1 alternating:4 symmetric:1 semantic:1 during:1 illustrative:1 coincides:2 criterion:2 generalized:9 instantaneous:1 ef:2 recently:1 common:1 behaves:1 ji:1 extend:2 interpretation:4 slight:1 belong:1 onedimensional:1 had:1 dot:1 similarity:1 surface:5 etc:1 closest:1 optimizing:2 verlag:1 binary:6 minimum:6 converge:2 determine:1 semi:1 stem:1 retrieval:1 ensuring:1 variant:4 regression:6 essentially:1 expectation:5 poisson:6 iteration:1 represent:1 tailored:2 want:1 rest:1 sr:1 pass:1 comment:2 subject:2 member:9 integer:3 restrict:1 andreas:1 idea:2 regarding:1 avenue:1 whether:1 pca:37 passing:1 hessian:2 cause:1 ignored:2 generally:1 useful:2 eigenvectors:2 simplest:1 schapire:2 canonical:2 notice:2 write:2 dasgupta:2 verified:2 sum:6 inverse:1 parameterized:2 place:1 family:26 almost:1 throughout:1 draw:2 decision:1 summarizes:1 bound:2 nonnegative:5 elucidated:1 annual:1 bv:1 infinity:1 constraint:3 argument:6 combination:1 conjugate:1 intimately:1 modification:1 happens:1 explained:1 projecting:1 outlier:1 indexing:1 glm:7 taken:4 remains:1 turn:2 discus:1 jolliffe:3 generalizes:1 gaussians:1 operation:3 permit:1 appropriate:2 generic:4 alternative:1 existence:1 thomas:1 denotes:2 binomial:1 assumes:1 include:1 build:1 approximating:1 hypercube:2 society:1 tusn:3 added:2 usual:1 forster:1 subspace:10 distance:22 link:4 mapped:1 simulated:1 argue:1 length:1 pointwise:1 relationship:2 index:1 illustration:1 minimizing:2 difficult:1 robert:1 fe:1 negative:6 ba:1 twenty:1 finite:1 inevitably:1 defining:1 extended:1 buja:1 ordinate:1 pair:1 optimized:3 componentwise:1 alternately:2 usually:6 summarize:1 program:2 royal:1 tance:1 power:4 suitable:1 critical:1 natural:10 hybrid:1 rely:1 representing:1 scheme:2 lk:1 prior:2 literature:1 swerve:1 acknowledgement:1 zh:1 relative:2 loss:26 expect:1 interesting:3 rameters:1 generation:1 row:5 keeping:1 side:1 allow:1 taking:1 fg:1 curve:5 commonly:1 projected:1 simplified:1 founded:1 approximate:3 implicitly:1 global:4 proximation:1 nelder:1 un:1 continuous:1 search:2 iterative:3 latent:1 table:2 nature:2 reasonably:1 ignoring:2 complex:1 domain:4 main:1 azoury:1 noise:2 edition:1 referred:1 sub:2 wish:1 exponential:29 lie:9 bishop:2 adding:1 preoccupation:1 supplement:1 suited:2 generalizing:1 fc:1 simply:4 saddle:1 expressed:1 katy:1 scalar:1 springer:1 acm:1 conditional:1 goal:2 presentation:1 identity:1 exposition:1 change:1 mccullagh:1 specifically:3 hyperplane:2 minization:2 principal:5 sanjoy:1 egf:1 maximumlikelihood:1 collins:1 overload:1
1,183
2,079
Improvisation and Learning Judy A. Franklin Computer Science Department Smith College Northampton, MA 01063 jfranklin@cs.smith.edu Abstract This article presents a 2-phase computational learning model and application. As a demonstration, a system has been built, called CHIME for Computer Human Interacting Musical Entity. In phase 1 of training, recurrent back-propagation trains the machine to reproduce 3 jazz melodies. The recurrent network is expanded and is further trained in phase 2 with a reinforcement learning algorithm and a critique produced by a set of basic rules for jazz improvisation. After each phase CHIME can interactively improvise with a human in real time. 1 Foundations Jazz improvisation is the creation of a jazz melody in real time. Charlie Parker, Dizzy Gillespie, Miles Davis, John Coltrane, Charles Mingus, Thelonious Monk, and Sonny Rollins et al. were the founders of bebop and post bop jazz [9] where drummers, bassists, and pianists keep the beat and maintain harmonic structure. Other players improvise over this structure and even take turns improvising for 4 bars at a time. This is called trading fours. Meanwhile, artificial neural networks have been used in computer music [4, 12]. In particular, the work of (Todd [11]) is the basis for phase 1 of CHIME, a novice machine improvisor that learns to trade fours. Firstly, a recurrent network is trained with back-propagation to play three jazz melodies by Sonny Rollins [1], as described in Section 2. Phase 2 uses actor-critic reinforcement learning and is described in Section 3. This section is on jazz basics. 1.1 Basics: Chords, the ii-V-I Chord Progression and Scales The harmonic structure mentioned above is a series of chords that may be reprated and that are often grouped into standard subsequences. A chord is a group of notes played simultaneously. In the chromatic scale, C-Db-D-Eb-E-F-Gb-G-Ab-A-Bb-B-C, notes are separated by a half step. A flat (b) note is a half step below the original note; a sharp (#) is a half above. Two half steps are a whole step. Two whole steps are a major third. Three half steps are a minor third. A major triad (chord) is the first or tonic note, then the note a major third up, then the note a minor third up. When F is the tonic, F major triad is F-A-C. A minor triad (chord) is the tonic  www.cs.smith.edu/?jfrankli then a minor third, then a major third. F minor triad is F-Ab-C. The diminished triad is the tonic, then a minor third, then a minor third. F diminished triad is F-Ab-Cb. An augmented triad is the tonic, then a major third, then a major third. The F augmented triad is F-A-Db. A third added to the top of a triad forms a seventh chord. A major triad plus a major third is the major seventh chord. F-A-C-E is the F major seventh chord (Fmaj7). A minor triad plus a minor third is a minor seventh chord. For F it is F-Ab-C-Eb (Fm7). A major triad plus a minor third is a dominant seventh chord. For F it is F-A-C-Eb (F7). These three types of chords are used heavily in jazz harmony. Notice that each note in the chromatic scales can be the tonic note for any of these types of chords. A scale, a subset of the chromatic scale, is characterized by note intervals. Let W be a whole step and H be a half. The chromatic scale is HHHHHHHHHHHH. The major scale or ionian mode is WWHWWWH. F major scale is F-G-A-Bb-C-D-E-F. The notes in a scale are degrees; E is the seventh degree of F major. The first, third, fifth, and seventh notes of a major scale are the major seventh chord. The first, third, fifth, and seventh notes of other modes produce the minor seventh and dominant seventh chords. Roman numerals represent scale degrees and their seventh chords. Upper case implies major or dominant seventh and lower case implies minor seventh [9]. The major seventh chord starting at the scale tonic is the I (one) chord. G is the second degree of F major, and G-Bb-D-F is Gm7, the ii chord, with respect to F. The ii-V-I progression is prevalent in jazz [9], and for F it is Gm7-C7-Fmaj7. The minor ii-V-i progression is obtained using diminished and augmented triads, their seventh chords, and the aeolian mode. Seventh chords can be extended by adding major or minor thirds, e.g. Fmaj9, Fmaj11, Fmaj13, Gm9, Gm11, and Gm13. Any extension can be raised or lowered by 1 step [9] to obtain, e.g. Fmaj7#11, C7#9, C7b9, C7#11. Most jazz compositions are either the 12 bar blues or sectional forms (e.g. ABAB, ABAC, or AABA) [8]. The 3 Rollins songs are 12 bar blues. ?Blue 7? has a simple blues form. In ?Solid? and ?Tenor Madness?, Rollins adds bebop variations to the blues form [1]. ii-V-I and VI-II-V-I progressions are added and G7+9 substitutes for the VI and F7+9 for the V (see section 1.2 below); the II-V in the last bar provides the turnaround to the I of the first bar to foster smooth repetition of the form. The result is at left and in Roman numeral notation Bb7 Bb7 Bb7 Bb7 I I I I Eb7 Eb7 Bb7 G7+9 IV IV I VI at right: Cm7 F7 Bb7 G7+9 C7 F7+9 ii V I VI II V 1.2 Scale Substitutions and Rules for Reinforcement Learning First note that the theory and rules derived in this subsection are used in Phase 2, to be described in Section 3. They are presented here since they derive from the jazz basics immediately preceding. One way a novice improvisor can play is to associate one scale with each chord and choose notes from that scale when the chord is presented in the musical score. Therefore, Rule 1 is that an improvisor may choose notes from a ?standard? scale associated with a chord. Next, the 4th degree of the scale is often avoided on a major or dominant seventh chord (Rule 3), unless the player can resolve its dissonance. The major 7th is an avoid note on a dominant seventh chord (Rule 4) since a dominant seventh chord and its scale contain the flat 7th, not the major 7th. Rule 2 contains many notes that can be added. A brief rationale is given next. The C7 in Gm7-C7-Fmaj7 may be replaced by a C7#11, a C7+ chord, or a C7b9b5 or C7alt chord [9]. The scales for C7+ and C7#11 make available the raised fourth (flat 5), and flat 6 (flat 13) for improvising. The C7b9b5 and C7alt (C7+9) chords and their scales make available the flat9, raised 9, flat5 and raised 5 [1]. These substitutions provide the notes of Rule 2. These rules (used in phase 2) are stated below, using for reinforcement values very bad (-1.0), bad (-0.5), a little bad (-0.25), ok (0.25), good (0.5), and very good (1.0). The rules are discussed further in Section 4. The Rule Set: 1) Any note in the scale associated with the chord is ok (except as noted in rule 3). 2) On a dominant seventh, hip notes 9, flat9, #9, #11, 13 or flat13 are very good. One hip note 2 times in a row is a little bad. 2 hip notes more than 2 times in a row is a little bad. 3) If the chord is a dominant seventh chord, a natural 4th note is bad. 4) If the chord is a dominant seventh chord, a natural 7th is very bad. 5) A rest is good unless it is held for more than 2 16th notes and then it is very bad. 6) Any note played longer than 1 beat (4 16th notes) is very bad. 7) If two consecutive notes match the human?s, that is good. 2 CHIME Phase 1 In Phase 1, supervised learning is used to train a recurrent network to reproduce the three Sonny Rollins melodies. 2.1 Network Details and Training The recurrent network?s output units are linear. The hidden units are nonlinear (logistic function). Todd [11] used a Jordan recurrent network [6] for classical melody learning and generation. In CHIME, a Jordan net is also used, with the addition of the chord as input (Figure 1. 24 of the 26 outputs are notes (2 chromatic octaves), the 25th is a rest, and the 26th indicates a new note. The output with the highest value above a threshold is the next note, including the rest output. The new note output if this is a new note, or if it indicates  is the same note being held for another time step ( note resolution). The 12 chord inputs (12 notes in a chromatic scale), are 1 or 0. A chord is represented as its first, third, fifth, and seventh notes and it ?wraps around? within the 12 inputs. E.g., the Fm7 chord F-Ab-C-Eb is represented as C, Eb, F, Ab or 100101001000. One plan input per song enables distinguishing between songs. The 26 context inputs use eligibility traces, giving the hidden units a decaying history of notes played. CHIME (as did Todd) uses teacher forcing [13], wherein the target outputs for the previous step are used as inputs (so erroneous outputs are not used as inputs). Todd used from 8 to 15 hidden units; CHIME uses 50. The learning rate is 0.075 (Todd used 0.05). The eligibility rate is 0.9 (Todd used 0.8). Differences in values perhaps reflect contrasting styles of the songs and available computing power. Todd used 15 output units and assumed a rest when all note units are ?turned off.? CHIME uses 24 output note units (2 octaves). Long rests in the Rollins tunes require a dedicated output unit for a rest. Without it, the note outputs learned to turn off all the time. Below are results of four representative experiments. In all experiments, 15,000 presentations of the songs were made. Each song has 192 16th note events. All songs are played at a fixed tempo. Weights are initialized to small random values. The squared error is the average squared error over one complete presentation of the song. ?Finessing? the network may improve these values. The songs are easily recognized however, and an exact match could impair the network?s ability to improvise. Figure 2 shows the results for ?Solid.? Experiment 1. Song: Blue Seven. Squared error starts at 185, decreases to 2.67. Experiment 2. Song: Tenor Madness. Squared error starts at 218, decreases to 1.05. Experiment 3. Song: Solid. Squared error starts at 184, decreases to 3.77. Experiment 4. Song: All three songs: Squared error starts at 185, decreases to 36. Figure 1: Jordan recurrent net with addition of chord input 2.2 Phase 1 Human Computer Interaction in Real Time In trading fours with the trained network, human note events are brought in via the MIDI interface [7]. Four bars of human notes are recorded then given, one note event at a time to the context inputs (replacing the recurrent inputs). The plan inputs are all 1. The chord inputs follow the ?Solid? form. The machine generates its four bars and they are played in real time. Then the human plays again, etc. An accompaniment (drums, bass, and piano), produced by Band-in-a-Box software (PG Music), keeps the beat and provides chords for the human. Figure 3 shows an interaction. The machine?s improvisations are in the second and fourth lines. In bar 5 the flat 9 of the Eb7 appears; the E. This note is used on the Eb7 and Bb7 chords by Rollins in ?Blue 7?, as a ?passing tone.? D is played in bar 5 on the Eb7. D is the natural 7 over Eb7 (with its flat 7) but is a note that Rollins uses heavily in all three songs, and once over the Eb7. It may be a response to the rest and the Bb played by the human in bar 1. D follows both a rest and a Bb in many places in ?Tenor Madness? and ?Solid.? In bar 6, the long G and the Ab (the third then fourth of Eb7) figure prominently in ?Solid.? At the beginning of bar 7 is the 2-note sequence Ab-E that appears in exactly the same place in the song ?Blue 7.? The focus of bars 7 and 8 is jumping between the 3rd and 4th of Bb7. At the end of bar 8 the machine plays the flat 9 (Ab) then the flat 3 (Bb), of G7+9. In bars 13-16 the tones are longer, as are the human?s in bars 9-12. The tones are the 5th, the root, the 3rd, the root, the flat 7, the 3rd, the 7th, and the raised fourth. Except for the last 2, these are chord tones. 3 CHIME Phase 2 In Phase 2, the network is expanded and trained by reinforcement learning to improvise according to the rules of Section 1.2 and using its knowledge of the Sonny Rollins songs. 3.1 The Expanded Network Figure 4 shows the phase 2 network. The same inputs plus 26 human inputs brings the total to 68. The weights obtained in phase 1 initialize this network. The plan and chord weights Figure 2: At left ?Solid? played by a human; at right the song reproduced by the ANN. are the same. The weights connecting context units to the hidden layer are halved. The same weights, halved, connect the 26 human inputs to the hidden layer. Each output unit gets the 100 hidden units? outputs as input. The original 50 weights are halved and used as initial values of the two sets of 50 hidden unit weights to the output unit. 3.2 SSR and Critic Algorithms Using actor-critic reinforcement learning ([2, 10, 13]), the actor chooses the next note to play. The critic receives a ?raw? reinforcement signal from the critique made by the rules of Section 1.2. For output j, the SSR (actor) computes mean      . A Gaussian distribution with mean  and standard deviation  chooses the output  . is generated, the critic modifies and produces  .  is further modified by a self-scaling algorithm that tracks, via moving average, the maximum and minimum reinforcement and uses them to scale the signal to produce   !#"$%'&(*)+",-%/.102"$3) . %2&4 ! 56 %2&4 ! 56 )78%'&49%'&4-5:)3;: < )7A@B%2&4 !56 )6A ",@C)D  %/.10=-56 %/.10=-56 )= %2.>0+9%/.10=!5:)?;: 3< )=E@B%2.10=!5F6 )6A ",@C)D  The goal is to make small gains in reinforcement more noticeable and to scale the values "A !%2.>0H" %2&4*)I "AJ and if between -1 and 1. If  G%'&4 , then    8%2.10 , then 8 !%2.>02"I%'&4K)+"  J7" . If JMLNLHO , the extremes of -1 and 1 are approached. The weight and standard deviation updates use  :  P!5F6 )=Q 4!5:)F6SRT !M"UK)DVKKWXVK    -5F6 )=8%'&49CY[Z3\;]%/.10+9_^  !5:)6  "I^ )`a!%'&4b"U%2.10F)3;cBY < < ;cOdLH^IL  If the difference between the max and min reinforcement stays large, over time   will increase (to a max of  Y[Z3\ ) and allow more exploration. When rmax-rmin is small, over time e will shrink (to a min of  Y ). The actor?s hidden units are updated using back   ) as ?error.? See [3, 5, 13] for more details on the propagation as before, using f`T!  "g SSR algorithm and its precursors. The critic inputs are the outputs of the hidden layer of the actor network; it ?piggy-backs? on the actor and uses its learned features (see Figure 5). This also alleviates the computational burden so it can run in real time. There are delays in reward, e.g. in that a note played too many times in a row may result in punishment, and if 2 notes in a row coincide with the human?s it is rewarded. If !;:5:) is the prediction of future reward [10] for state x at time t, Figure 3: Phase1 trading 4 bars: 4 human, 4 machine, 4 human, 4 machine  -56 )= e!5:)F6,^  ` !-5F6 )3;:5:) "2--5:)3;:5:) for O The critic is a linear function of its inputs:  !-5:)3;]5:)T updated incrementally using the value of  : LHY ^  L .    ! 5:)D !5:) The weights are  -5F6 )=Q -5:)F6  V4*WVK  is in effect an error signal, a difference between consecutive predicted rewards [10]. The critic also uses eligibility traces of the inputs, so !5:) is actually   !5:)a  !5:)76  \ !5=" ) where  \ -5:)     !5:) . While this is all experimental, initial results show that the system with both the self-scalar and the critic performs better than with just one or without either one. A more systematic study is planned. 4 Results and Comments Recall the rules of Section 1.2. Rules 1-4 are based on discussions with John Payne, a professional jazz musician and instructor of 25 years 1 . The rules by no means encompass all of jazz theory or practice but are a starting point. The notes in rule 2 were cast as good in a ?hip? situation. The notion of hip requires human sophistication so for now these notes are reinforced if played sporadically on the dominant seventh. Rule 5 was added to discourage not playing any notes. Rules 5 and 6 focus on not allowing an output of one note for too long. Each chord is assigned a scale for rule 1. C is limited, O O $L egL O , providing stability, and deliberate action uncertainty so different notes are played, for the same network state. Generally the goal of reinforcement learning is to find the best action for a given state, with uncertainty used for further exploration. Here, reinforcement learning finds the best set of actions for a given state. In a typical example using the phase 1 network prior to phase 2 improvement, the average reinforcement value according to the rule set is -.37 (on a scale from -1 to 1). After Phase 2, the average reinforcement value is .28 after 30-100 off-line presentations of the human solo of 1800 note events. Figure 6 shows 12 bars of a human solo and 12 bars of a machine solo. The note durations 1 The rules are not meant to represent John Payne Figure 4: Recurrent reinforcement learning network with human input used in phase 2. Figure 5: Phase 2 network with critic ?piggy-backing? on hidden layer. are shortened, reflecting the rules to prevent settling onto one note. The machine plays chord tones, such as Bb and D in bars 1 and 2. The high G is the 13 of Bb7, a hip note. In bars 3 and 4 it plays C sharp, a hip note (the #9 of Bb7) and high G. These notes are played in bars 9 and 11 on Bb7. In bars 5 and 6 the 9 and 13 (F and C) are played on Eb7. The natural 7 (D) reflects its heavy use in Rollins? melodies. Hip notes show up in bar 9 on Cm7: the 13 (G) and the 9 (D). In bar 11 G is played on G7+9 as is the hip flat 9 (the Ab). In bar 12, the Eb (flat 7 chord tone) is played on the F7+9. In bars 2, 4, 7, 9, and 10 the machine starts at the G at the top of the staff and descends through several chord tones, producing a recurring motif, an artifact of a ?good? jazz solo. The phase 2 network has been used to interact with a human in real time while still learning. It keeps its recurrence since the human has a separate set of inputs. A limitation to be addressed for CHIME is to move beyond one chord at a time. To achieve this, it must use more context, over more time. There are plenty of improvisation rules for chord progressions [8]. Because CHIME employs reinforcement learning, it has a stochastic element that allows it to play ?outside the chord changes.? A research topic is to understand how to enable it to do this more pointedly. Figure 6: At left, 12 bars of human solo. At right, 12 bars the machine plays in response. References [1] J. Aebersold. You can play Sonny Rollins. A New Approach to Jazz Improvisation Vol 8. Jamey Aebersold, New Albany, IND., 1976. [2] A. G. Barto, R. S. Sutton, and C. W. Anderson. Neuronlike adaptive element that can solve difficult learning control problems. IEEE Transactions on Systems, Man, and Cybernetics, SMC13:834?846, 1983. [3] H. Benbrahim and J. Franklin. Biped walking using reinforcement learning. Robotics and Autonomous Systems, 22:283?302, 1997. [4] N. Griffith and P. Todd. Musical Networks: Parallel Distributed Perception and Performance. MIT Press, Cambridge MA, 1999. [5] V. Gullapalli, J. Franklin, and H. Benbrahim. Acquiring robot skills via reinforcement learning. IEEE Control Systems Magazine, 1994. [6] M. Jordan. Attractor dynamics and parallelism in a connectionist sequential machine. In Proceedings of the Eighth Annual Conference of the Cognitive Science Society, 1986. [7] P. Messick. Maximum MIDI. Manning Publications, Greenwich, CT, 1988. [8] S. Reeves. Creative Jazz Improvisation. 2nd Ed. Prentice Hall, Upper Saddle River NJ, 1995. [9] M. A. Sabatella. Whole Approach to Jazz Improvisation. A.D.G. Productions, Lawndale CA, 1996. [10] R. Sutton. Learning to predict by the methods of temporal differences. Machine Learning, 3:9? 44, 1988. [11] P. M. Todd. A connectionist approach to algorithmic composition. In P. M. Todd and e. D. Loy, editors, Music and Connectionism. MIT Press, Cambridge MA, 1991. [12] P. M. Todd and e. D. Loy. Music and Connectionism. MIT Press, Cambridge, MA, 1991. [13] R. J. Williams. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning, 8:229?256, 1992.
2079 |@word nd:1 pg:1 solid:7 initial:2 substitution:2 series:1 score:1 contains:1 accompaniment:1 franklin:3 must:1 john:3 enables:1 update:1 half:6 tone:7 monk:1 beginning:1 smith:3 provides:2 firstly:1 resolve:1 little:3 precursor:1 notation:1 benbrahim:2 rmax:1 contrasting:1 nj:1 temporal:1 exactly:1 control:2 unit:14 producing:1 before:1 todd:11 sutton:2 shortened:1 critique:2 improvisation:8 plus:4 eb:6 limited:1 g7:5 practice:1 instructor:1 griffith:1 get:1 onto:1 prentice:1 context:4 www:1 musician:1 modifies:1 williams:1 starting:2 duration:1 resolution:1 immediately:1 rule:25 stability:1 notion:1 variation:1 autonomous:1 updated:2 target:1 play:10 heavily:2 magazine:1 exact:1 distinguishing:1 us:8 associate:1 element:2 walking:1 cy:1 triad:13 bass:1 trade:1 highest:1 chord:52 decrease:4 mentioned:1 reward:3 dynamic:1 trained:4 creation:1 basis:1 easily:1 represented:2 train:2 separated:1 artificial:1 approached:1 outside:1 solve:1 ability:1 reproduced:1 sequence:1 net:2 interaction:2 turned:1 payne:2 alleviates:1 achieve:1 aebersold:2 produce:3 derive:1 recurrent:9 minor:15 noticeable:1 descends:1 c:2 predicted:1 trading:3 implies:2 stochastic:1 exploration:2 human:23 enable:1 melody:6 numeral:2 require:1 connectionism:2 extension:1 around:1 hall:1 cb:1 algorithmic:1 predict:1 chime:11 major:24 consecutive:2 f7:5 albany:1 jazz:17 harmony:1 grouped:1 repetition:1 reflects:1 brought:1 mit:3 j7:1 gaussian:1 modified:1 avoid:1 chromatic:6 barto:1 publication:1 derived:1 focus:2 vk:1 improvement:1 prevalent:1 indicates:2 motif:1 hidden:10 reproduce:2 backing:1 plan:3 raised:5 initialize:1 once:1 dissonance:1 plenty:1 future:1 connectionist:3 roman:2 employ:1 simultaneously:1 replaced:1 phase:21 attractor:1 maintain:1 ab:10 neuronlike:1 extreme:1 held:2 solo:5 jumping:1 unless:2 iv:2 initialized:1 hip:9 planned:1 deviation:2 subset:1 delay:1 seventh:25 too:2 connect:1 teacher:1 chooses:2 punishment:1 river:1 stay:1 v4:1 off:3 systematic:1 connecting:1 again:1 squared:6 reflect:1 interactively:1 recorded:1 choose:2 cognitive:1 style:1 f6:7 rollins:11 vi:4 root:2 sporadically:1 start:5 decaying:1 parallel:1 il:1 musical:3 reinforced:1 lhy:1 raw:1 produced:2 cybernetics:1 history:1 ed:1 c7:11 associated:2 gain:1 recall:1 subsection:1 knowledge:1 actually:1 back:4 reflecting:1 appears:2 ok:2 supervised:1 follow:1 wherein:1 response:2 box:1 shrink:1 anderson:1 just:1 receives:1 replacing:1 nonlinear:1 propagation:3 incrementally:1 dvk:1 ionian:1 brings:1 aj:1 artifact:1 perhaps:1 mode:3 logistic:1 effect:1 contain:1 assigned:1 mile:1 ind:1 self:2 eligibility:3 recurrence:1 davis:1 noted:1 octave:2 complete:1 performs:1 dedicated:1 interface:1 harmonic:2 charles:1 discussed:1 composition:2 cambridge:3 rd:3 reef:1 biped:1 lowered:1 improvise:4 actor:7 longer:2 moving:1 robot:1 etc:1 add:1 dominant:10 halved:3 forcing:1 rewarded:1 minimum:1 preceding:1 staff:1 recognized:1 signal:3 ii:9 encompass:1 smooth:1 match:2 characterized:1 long:3 post:1 prediction:1 basic:4 represent:2 robotics:1 addition:2 interval:1 addressed:1 bop:1 rest:8 comment:1 db:2 jordan:4 drum:1 gullapalli:1 gb:1 song:18 passing:1 action:3 generally:1 tune:1 band:1 deliberate:1 notice:1 per:1 track:1 blue:8 vol:1 group:1 four:6 threshold:1 prevent:1 year:1 run:1 fourth:4 uncertainty:2 you:1 place:2 improvising:2 scaling:1 layer:4 ct:1 played:15 annual:1 rmin:1 flat:12 software:1 generates:1 min:2 expanded:3 department:1 according:2 creative:1 manning:1 founder:1 turn:2 end:1 available:3 progression:5 tempo:1 professional:1 substitute:1 original:2 top:2 charlie:1 music:4 giving:1 classical:1 society:1 move:1 added:4 abab:1 gradient:1 wrap:1 separate:1 entity:1 seven:1 topic:1 z3:2 providing:1 demonstration:1 loy:2 difficult:1 trace:2 stated:1 allowing:1 upper:2 beat:3 situation:1 extended:1 tonic:7 ssr:3 interacting:1 sharp:2 cast:1 learned:2 impair:1 bar:29 recurring:1 below:4 beyond:1 perception:1 parallelism:1 eighth:1 built:1 including:1 max:2 gillespie:1 power:1 event:4 natural:4 settling:1 improve:1 brief:1 pianist:1 prior:1 piano:1 rationale:1 generation:1 limitation:1 foundation:1 degree:5 article:1 foster:1 editor:1 playing:1 critic:10 heavy:1 production:1 row:4 aaba:1 gl:1 last:2 allow:1 understand:1 fifth:3 distributed:1 computes:1 made:2 reinforcement:19 coincide:1 avoided:1 adaptive:1 novice:2 transaction:1 bb:7 skill:1 midi:2 keep:3 assumed:1 subsequence:1 tenor:3 ca:1 interact:1 meanwhile:1 discourage:1 did:1 whole:4 augmented:3 representative:1 parker:1 judy:1 prominently:1 third:19 learns:1 erroneous:1 bad:9 burden:1 adding:1 sequential:1 madness:3 sophistication:1 saddle:1 sectional:1 scalar:1 acquiring:1 ma:4 piggy:2 goal:2 presentation:3 ann:1 man:1 change:1 diminished:3 abac:1 except:2 typical:1 called:2 total:1 experimental:1 player:2 college:1 meant:1 gm13:1
1,184
208
550 Ackley and Littman Generalization and scaling in reinforcement learning David H. Ackley Michael L. Littman Cognitive Science Research Group Bellcore Morristown, NJ 07960 ABSTRACT In associative reinforcement learning, an environment generates input vectors, a learning system generates possible output vectors, and a reinforcement function computes feedback signals from the input-output pairs. The task is to discover and remember input-output pairs that generate rewards. Especially difficult cases occur when rewards are rare, since the expected time for any algorithm can grow exponentially with the size of the problem. Nonetheless, if a reinforcement function possesses regularities, and a learning algorithm exploits them, learning time can be reduced below that of non-generalizing algorithms. This paper describes a neural network algorithm called complementary reinforcement back-propagation (CRBP), and reports simulation results on problems designed to offer differing opportunities for generalization. 1 REINFORCEMENT LEARNING REQUIRES SEARCH Reinforcement learning (Sutton, 1984; Barto & Anandan, 1985; Ackley, 1988; Allen, 1989) requires more from a learner than does the more familiar supervised learning paradigm. Supervised learning supplies the correct answers to the learner, whereas reinforcement learning requires the learner to discover the correct outputs before they can be stored. The reinforcement paradigm divides neatly into search and learning aspects: When rewarded the system makes internal adjustments to learn the discovered input-output pair; when punished the system makes internal adjustments to search elsewhere. Generalization and Scaling in Reinforcement Learning 1.1 MAKING REINFORCEMENT INTO ERROR Following work by Anderson (1986) and Williams (1988), we extend the backpropagation algorithm to associative reinforcement learning. Start with a "garden variety" backpropagation network: A vector i of n binary input units propagates through zero or more layers of hidden units, ultimately reaching a vector 8 of m sigmoid units, each taking continuous values in the range (0,1). Interpret each 8j as the probability that an associated random bit OJ takes on value 1. Let us call the continuous, deterministic vector 8 the search vector to distinguish it from the stochastic binary output vector o. Given an input vector, we forward propagate to produce a search vector 8, and then perform m independent Bernoulli trials to produce an output vector o. The i - 0 pair is evaluated by the reinforcement function and reward or punishment ensues. Suppose reward occurs. We therefore want to make 0 more likely given i. Backpropagation will do just that if we take 0 as the desired target to produce an error vector (0 - 8) and adjust weights normally. Now suppose punishment occurs, indicating 0 does not correspond with i. By choice of error vector, backpropagation allows us to push the search vector in any direction; which way should we go? In absence of problem-specific information, we cannot pick an appropriate direction with certainty. Any decision will involve assumptions. A very minimal "don't be like 0" assumption-employed in Anderson (1986), Williams (1988), and Ackley (1989)-pushes s directly away from 0 by taking (8 - 0) as the error vector. A slightly stronger "be like not-o" assumption-employed in Barto & Anandan (1985) and Ackley (1987)-pushes s directly toward the complement of 0 by taking ((1 - 0) - 8) as the error vector. Although the two approaches always agree on the signs of the error terms, they differ in magnitudes. In this work, we explore the second possibility, embodied in an algorithm called complementary reinforcement back-propagation ( CRBP). Figure 1 summarizes the CRBP algorithm. The algorithm in the figure reflects three modifications to the basic approach just sketched. First, in step 2, instead of using the 8j'S directly as probabilities, we found it advantageous to "stretch" the values using a parameter v. When v < 1, it is not necessary for the 8i'S to reach zero or one to produce a deterministic output. Second, in step 6, we found it important to use a smaller learning rate for punishment compared to reward. Third, consider step 7: Another forward propagation is performed, another stochastic binary output vector 0* is generated (using the procedure from step 2), and 0* is compared to o. If they are identical and punishment occurred, or if they are different and reward occurred, then another error vector is generated and another weight update is performed. This loop continues until a different output is generated (in the case of failure) or until the original output is regenerated (in the case of success). This modification improved performance significantly, and added only a small percentage to the total number of weight updates performed. 551 552 Ackley and Littman O. Build a back propagation network with input dimensionality n and output dimensionality m. Let t = 0 and te = O. 1. Pick random i E 2n and forward propagate to produce a/s. 2. Generate a binary output vector o. Given a uniform random variable ~ E [0,1] and parameter 0 < v < 1, OJ = {1, 0, if(sj - !)/v+! ~ ~j otherwise. 3. Compute reinforcement r = f(i,o). Increment t. If r < 0, let te = t. 4. Generate output errors ej. If r > 0, let tj = OJ, otherwise let tj = 1- OJ. Let ej = (tj - sj)sj(l- Sj). 5. Backpropagate errors. 6. Update weights. 1:::..Wjk = 1]ekSj, using 1] = 1]+ if r ~ 0, and 1] = 1]- otherwise, with parameters 1]+,1]- > o. 7. Forward propagate again to produce new Sj's. Generate temporary output vector 0*. If (r > 0 and 0* #- 0) or (r < 0 and 0* = 0), go to 4. 8. If te ~ t, exit returning te, else go to 1. Figure 1: Complementary Reinforcement Back Propagation-CRBP 2 ON-LINE GENERALIZATION When there are many possible outputs and correct pairings are rare, the computational cost associated with the search for the correct answers can be profound. The search for correct pairings will be accelerated if the search strategy can effectively generalize the reinforcement received on one input to others. The speed of an algorithm on a given problem relative to non-generalizing algorithms provides a measure of generalization that we call on-line generalization. O. Let z be an array of length 2n. Set the z[i] to random numbers from 0 to 2m - 1. Let t = te = O. 1. Pick a random input i E 2n. 2. Compute reinforcement r = f(i, z[i]). Increment t. 3. If r < 0 let z[i] = (z[i] + 1) mod 2m , and let te = t. 4. If te <t:: t exit returning t e, else go to 1. Figure 2: The Table Lookup Reference Algorithm Tref(f, n, m) Consider the table-lookup algorithm Tref(f, n, m) summarized in Figure 2. In this algorithm, a separate storage location is used for each possible input. This prevents the memorization of one i - 0 pair from interfering with any other. Similarly, the selection of a candidate output vector depends only on the slot of the table corresponding to the given input. The learning speed of T ref depends only on the input and output dimensionalities and the number of correct outputs associated Generalization and Scaling in Reinforcement Learning with each input. When a problem possesses n input bits and n output bits, and there is only one correct output vector for each input vector, Tre{ runs in about 4n time (counting each input-output judgment as one.) In such cases one expects to take at least 2n - 1 just to find one correct i - 0 pair, so exponential time cannot be avoided without a priori information. How does a generalizing algorithm such as CRBP compare to Trer? 3 SIMULATIONS ON SCALABLE PROBLEMS We have tested CRBP on several simple problems designed to offer varying degrees and types of generalization. In all of the simulations in this section, the following details apply: Input and output bit counts are equal (n). Parameters are dependent on n but independent of the reinforcement function f. '7+ is hand-picked for each n,l 11- = 11+/10 and II = 0.5. All data points are medians of five runs. The stopping criterion te ~ t is interpreted as te +max(2000, 2n+l) < t. The fit lines in the figures are least squares solutions to a x bn , to two significant digits. As a notational convenience, let c = ~ 3.1 n E ij ;=1 - the fraction of ones in the input. n-MAJORlTY Consider this "majority rules" problem: [if c > ~ then 0 = In else 0 = on]. The i-o mapping is many-to-l. This problem provides an opportunity for what Anderson (1986) called "output generalization": since there are only two correct output states, every pair of output bits are completely correlated in the cases when reward occurs. G) 'iii u rn C) 0 ::::. G) E ; 10 7 10 6 10 5 10 4 x Table D CRBP n-n-n + CRBP n-n 10 3 10 2 10 1 10 0 0 1 2 3 456 78 91011121314 n Figure 3: The n-majority problem Figure 3 displays the simulation results. Note that although Trer is faster than CRBP at small values of n, CRBP's slower growth rate (1.6n vs 4.2n ) allows it to cross over and begin outperforming Trer at about 6 bits. Note also--in violation of 1 For n = 1 to 12. we used '1+ 0.219. 0.170. 0.121}. = {2.000. 1.550. 1.130.0.979.0.783.0.709.0.623.0.525.0.280. 553 554 Ackley and Littman some conventional wisdom-that although n-majority is a linearly separable problem, the performance of CRBP with hidden units is better than without. Hidden units can be helpful--even on linearly separable problems-when there are opportunities for output generalization. 3.2 n-COPY AND THE 2k -ATTRACTORS FAMILY As a second example, consider the n-copy problem: [0 = i]. The i-o mapping is now 1-1, and the values of output bits in rewarding states are completely uncorrelated, but the value of each output bit is completely correlated with the value of the corresponding input bit. Figure 4 displays the simulation results. Once again, at G) 'ii tA Q 0 ::::. G) - .5 10 7 10 6 10 5 10 4 x 150*2.0I\n D 10 3 10 2 12*2.2I\n + Table CRBP n-n-n CRBP n-n 10 1 10 0 0 1 2 3 4 5 6 7 8 9 10 1112 n Figure 4: The n-copy problem low values of n, Trer is faster, but CRBP rapidly overtakes Trer as n increases. In n-copy, unlike n-majority, CRBP performs better without hidden units. The n-majority and n-copy problems are extreme cases of a spectrum. n-majority can be viewed as a "2-attractors" problem in that there are only two correct outputs-all zeros and all ones-and the correct output is the one that i is closer to in hamming distance. By dividing the input and output bits into two groups and performing the majority function independently on each group, one generates a "4-aUractors" problem. In general, by dividing the input and output bits into 1 ~ Ie ~ n groups, one generates a "2i:-attractors" problem. When Ie = 1, nmajority results, and when Ie n, n-copy results. = Figure 5 displays simulation results on the n = 8-bit problems generated when Ie is varied from 1 to n. The advantage of hidden units for low values of Ie is evident, as is the advantage of "shortcut connections" (direct input-to-output weights) for larger values of Ie. Note also that combination of both hidden units and shortcut connections performs better than either alone. Generalization and Scaling in Reinforcement Learning 105~--------------------------------~ CASP 8-10-8 -+- CASP 8-8 .... CASP 8-10-Sls -0- ... Table 3 2 1 5 4 7 6 8 k Figure 5: The 21:- attractors family at n = 8 3.3 n-EXCLUDED MIDDLE All of the functions considered so far have been linearly separable. Consider this "folded majority" function: [if < c < then 0 on else 0 In]. Now, like n-majority, there are only two rewarding output states, but the determination of which output state is correct is not linearly separable in the input space. When n = 2, the n-excluded middle problem yields the EQV (i.e., the complement of XOR) function, but whereas functions such as n-parity [if nc is even then 0 on else 0 = In] get more non-linear with increasing n, n-excluded middle does not. i i = = = 107~------------------------------~~ - 10 6 10 5 D) 10 4 10 3 I) 'ii u f) .2 I) E ::: x c 17oo*1.6"n Table CRSP n-n-n/s 10 2 10 1 10 0 0 1 2 3 4 5 6 7 8 9 10 1112 n Figure 6: The n-excluded middle problem Figure 6 displays the simulation results. CRBP is slowed somewhat compared to the linearly separable problems, yielding a higher "cross over point" of about 8 bits. 555 556 Ackley and Littman 4 STRUCTURING DEGENERATE OUTPUT SPACES All of the scaling problems in the previous section are designed so that there is a single correct output for each possible input. This allows for difficult problems even at small sizes, but it rules out an important aspect of generalizing algorithms for associative reinforcement learning: If there are multiple satisfactory outputs for given inputs, a generalizing algorithm may impose structure on the mapping it produces. We have two demonstrations of this effect, "Bit Count" and "Inverse Arithmetic." The Bit Count problem simply states that the number of I-bits in the output should equal the number of I-bits in the input. When n = 9, Tref rapidly finds solutions involving hundreds of different output patterns. CRBP is slower--especially with relatively few hidden units-but it regularly finds solutions involving just 10 output patterns that form a sequence from 09 to 19 with one bit changing per step. 0+Ox4=0 1+0x4=1 2+0x4=2 3+0x4=3 0+2x4=8 1+2x4=9 2 + 2 x 4 = 10 3+2x4=11 4+0x4=4 4+ 2 x 4 = 5+0x4=5 5 + 2 x 4 = 6+0x4=6 6 + 2 x 4 = 7+0x4=7 7 + 2 x 4 = 12 13 14 15 2+2-4=0 2+2+4=8 3+2-4=1 3+2+4=9 2+2+4=2 2 + 2 x 4 = 10 3+2+4=3 3+2x4=1l 6+2-4=4 7+2-4=5 6+2+4=6 7+2-.;-4=7 6+ 7+ 6+ 7+ 2+ 4 = 2+ 4 = 2x4= 2x4= 0+4 x 4 = 16 0+6 x 4 = 1+4x4=17 1 + 6 x 4 = 2 + 4 x 4 = 18 2 + 6 x 4 = 3 +4 x 4 = 19 3 + 6 x 4 = 24 25 26 27 4+4 5+ 4 6+ 4 7+ 4 = = = = 28 29 30 31 24 25 26 27 x x x x 4= 4= 4= 4= 6+ 6 + 4 = 7+6+4= 2+ 4 x 4 = 3+ 4 x 4= 12 4 x 4 + 13 5 + 4 x 14 6 + 4 x 15 7 +4 x 4= 4= 4 4= = 20 4 + 6 x 21 5 + 6 x 22 6 + 6 x 23 7 + 6 x 4 4 4 4 16 17 18 19 0+6 x 1+ 6 x 2+ 6x 3+ 6x 4= 4= 4= 4= 20 21 22 23 4+ 5+ 6+ 7+ 4 = 28 4 = 29 4 30 4 = 31 6 6 6 6 x x x x = Figure 7: Sample CRBP solutions to Inverse Arithmetic The Inverse Arithmetic problem can be summarized as follows: Given i E 25 , find :1:, y, z E 23 and 0, <> E {+(OO)' -(01)' X (10)' +(11)} such that :I: oy<>z = i. In all there are 13 bits of output, interpreted as three 3-bit binary numbers and two 2-bit operators, and the task is to pick an output that evaluates to the given 5-bit binary input under the usual rules: operator precedence, left-right evaluation, integer division, and division by zero fails. As shown in Figure 7, CRBP sometimes solves this problem essentially by discovering positional notation, and sometimes produces less-globally structured solutions, particularly as outputs for lower-valued i's, which have a wider range of solutions. Generalization and Scaling in Reinforcement Learning 5 CONCLUSIONS Some basic concepts of supervised learning appear in different guises when the paradigm of reinforcement learning is applied to large output spaces. Rather than a "learning phase" followed by a "generalization test," in reinforcement learning the search problem is a generalization test, performed simultaneously with learning. Information is put to work as soon as it is acquired. The problem of of "overfitting" or "learning the noise" seems to be less of an issue, since learning stops automatically when consistent success is reached. In experiments not reported here we gradually increased the number of hidden units on the 8-bit copy problem from 8 to 25 without observing the performance decline associated with "too many free parameters." The 2 k -attractors (and 2 k -folds-generalizing Excluded Middle) families provide a starter set of sample problems with easily understood and distinctly different extreme cases. In degenerate output spaces, generalization decisions can be seen directly in the discovered mapping. Network analysis is not required to "see how the net does it." The possibility of ultimately generating useful new knowledge via reinforcement learning algorithms cannot be ruled out. References Ackley, D.H. (1987) A connectionist machine for genetic hillclimbing. Boston, MA: Kluwer Academic Press. Ackley, D.H. (1989) Associative learning via inhibitory search. In D.S. Touretzky (ed.), Advances in Neural Information Processing Systems 1, 20-28. San Mateo, CA: Morgan Kaufmann. Allen, R.B. (1989) Developing agent models with a neural reinforcement technique. IEEE Systems, Man, and Cybernetics Conference. Cambridge, MA. Anderson, C.W. (1986) Learning and problem solving with multilayer connectionist systems. University of Mass. Ph.D. dissertation. COINS TR 86-50. Amherst, MA. Barto, A.G. (1985) Learning by statistical cooperation of self-interested neuron-like computing elements. Human Neurobiology, 4:229-256. Barto, A.G., & Anandan, P. (1985) Pattern recognizing stochastic learning automata. IEEE Transactions on Systems, Man, and Cybernetics, 15, 360-374. Rumelhart, D.E., Hinton, G.E., & Williams, R.J. (1986) Learning representations by backpropagating errors. Nature, 323, 533-536. Sutton, R.S. (1984) Temporal credit assignment in reinforcement learning. University of Mass. Ph.D. dissertation. COINS TR 84-2. Amherst, MA. Williams, R.J. (1988) Toward a theory of reinforcement-learning connectionist systems. College of Computer Science of Northeastern University Technical Report NU-CCS-88-3. Boston, MA. 557
208 |@word trial:1 middle:5 seems:1 stronger:1 advantageous:1 simulation:7 propagate:3 bn:1 pick:4 tr:2 genetic:1 designed:3 update:3 v:1 alone:1 discovering:1 dissertation:2 provides:2 location:1 casp:3 five:1 direct:1 supply:1 profound:1 pairing:2 acquired:1 expected:1 globally:1 automatically:1 increasing:1 begin:1 discover:2 notation:1 mass:2 what:1 interpreted:2 differing:1 nj:1 certainty:1 remember:1 every:1 temporal:1 morristown:1 growth:1 returning:2 unit:10 normally:1 appear:1 before:1 understood:1 sutton:2 mateo:1 range:2 backpropagation:4 digit:1 procedure:1 significantly:1 get:1 cannot:3 convenience:1 selection:1 operator:2 storage:1 put:1 memorization:1 conventional:1 deterministic:2 williams:4 go:4 independently:1 automaton:1 rule:3 array:1 increment:2 target:1 suppose:2 element:1 rumelhart:1 particularly:1 continues:1 ackley:10 environment:1 reward:7 littman:5 ultimately:2 solving:1 division:2 exit:2 learner:3 completely:3 easily:1 larger:1 valued:1 otherwise:3 associative:4 advantage:2 sequence:1 net:1 loop:1 rapidly:2 degenerate:2 wjk:1 regularity:1 produce:8 generating:1 wider:1 oo:2 ij:1 received:1 solves:1 dividing:2 differ:1 direction:2 correct:13 stochastic:3 human:1 generalization:15 precedence:1 stretch:1 considered:1 credit:1 mapping:4 punished:1 reflects:1 always:1 reaching:1 rather:1 ej:2 varying:1 barto:4 structuring:1 notational:1 bernoulli:1 helpful:1 dependent:1 stopping:1 hidden:8 interested:1 sketched:1 issue:1 bellcore:1 priori:1 equal:2 once:1 identical:1 x4:14 regenerated:1 report:2 others:1 connectionist:3 few:1 simultaneously:1 familiar:1 phase:1 attractor:5 possibility:2 evaluation:1 adjust:1 violation:1 extreme:2 yielding:1 tj:3 closer:1 necessary:1 divide:1 desired:1 ruled:1 minimal:1 increased:1 assignment:1 cost:1 rare:2 expects:1 uniform:1 hundred:1 recognizing:1 too:1 stored:1 ensues:1 reported:1 answer:2 punishment:4 amherst:2 ie:6 rewarding:2 michael:1 again:2 cognitive:1 lookup:2 summarized:2 depends:2 performed:4 picked:1 observing:1 reached:1 start:1 square:1 xor:1 kaufmann:1 correspond:1 judgment:1 wisdom:1 yield:1 generalize:1 cc:1 cybernetics:2 reach:1 touretzky:1 ed:1 failure:1 evaluates:1 nonetheless:1 associated:4 hamming:1 stop:1 knowledge:1 dimensionality:3 back:4 ta:1 higher:1 supervised:3 improved:1 evaluated:1 anderson:4 just:4 until:2 hand:1 tre:1 propagation:5 effect:1 concept:1 excluded:5 satisfactory:1 self:1 backpropagating:1 criterion:1 evident:1 performs:2 allen:2 sigmoid:1 exponentially:1 extend:1 occurred:2 kluwer:1 interpret:1 significant:1 cambridge:1 similarly:1 neatly:1 rewarded:1 binary:6 success:2 outperforming:1 seen:1 morgan:1 anandan:3 somewhat:1 impose:1 employed:2 paradigm:3 signal:1 arithmetic:3 ii:3 multiple:1 technical:1 faster:2 determination:1 academic:1 offer:2 cross:2 involving:2 basic:2 scalable:1 multilayer:1 essentially:1 sometimes:2 whereas:2 want:1 else:5 grow:1 median:1 unlike:1 posse:2 regularly:1 mod:1 call:2 integer:1 counting:1 iii:1 variety:1 fit:1 decline:1 useful:1 involve:1 ph:2 reduced:1 generate:4 sl:1 percentage:1 inhibitory:1 sign:1 per:1 group:4 changing:1 fraction:1 run:2 inverse:3 family:3 decision:2 summarizes:1 scaling:6 bit:23 layer:1 followed:1 distinguish:1 display:4 fold:1 occur:1 generates:4 aspect:2 speed:2 performing:1 separable:5 relatively:1 tref:3 structured:1 developing:1 combination:1 describes:1 slightly:1 smaller:1 making:1 modification:2 slowed:1 gradually:1 agree:1 count:3 apply:1 away:1 appropriate:1 coin:2 slower:2 original:1 opportunity:3 exploit:1 especially:2 build:1 added:1 occurs:3 strategy:1 usual:1 distance:1 separate:1 majority:9 toward:2 length:1 demonstration:1 nc:1 difficult:2 perform:1 neuron:1 neurobiology:1 hinton:1 discovered:2 rn:1 overtakes:1 varied:1 david:1 complement:2 pair:7 required:1 connection:2 temporary:1 nu:1 starter:1 below:1 pattern:3 oj:4 max:1 garden:1 embodied:1 relative:1 oy:1 degree:1 agent:1 consistent:1 propagates:1 uncorrelated:1 interfering:1 elsewhere:1 cooperation:1 parity:1 copy:7 soon:1 free:1 taking:3 distinctly:1 feedback:1 eqv:1 computes:1 forward:4 reinforcement:29 san:1 avoided:1 far:1 transaction:1 sj:5 overfitting:1 don:1 spectrum:1 search:11 continuous:2 table:7 learn:1 nature:1 ca:1 linearly:5 noise:1 complementary:3 ref:1 fails:1 guise:1 exponential:1 candidate:1 third:1 northeastern:1 specific:1 effectively:1 magnitude:1 te:9 push:3 boston:2 backpropagate:1 generalizing:6 simply:1 likely:1 explore:1 positional:1 hillclimbing:1 prevents:1 adjustment:2 ma:5 slot:1 viewed:1 absence:1 shortcut:2 man:2 folded:1 called:3 total:1 indicating:1 college:1 internal:2 accelerated:1 tested:1 correlated:2
1,185
2,080
The Emergence of Multiple Movement Units in the Presence of Noise and Feedback Delay Michael Kositsky Andrew G. Barto Department of Computer Science University of Massachusetts Amherst, MA 01003-4610 kositsky,barto @cs.umass.edu  Abstract Tangential hand velocity profiles of rapid human arm movements often appear as sequences of several bell-shaped acceleration-deceleration phases called submovements or movement units. This suggests how the nervous system might efficiently control a motor plant in the presence of noise and feedback delay. Another critical observation is that stochasticity in a motor control problem makes the optimal control policy essentially different from the optimal control policy for the deterministic case. We use a simplified dynamic model of an arm and address rapid aimed arm movements. We use reinforcement learning as a tool to approximate the optimal policy in the presence of noise and feedback delay. Using a simplified model we show that multiple submovements emerge as an optimal policy in the presence of noise and feedback delay. The optimal policy in this situation is to drive the arm?s end point close to the target by one fast submovement and then apply a few slow submovements to accurately drive the arm?s end point into the target region. In our simulations, the controller sometimes generates corrective submovements before the initial fast submovement is completed, much like the predictive corrections observed in a number of psychophysical experiments. 1 Introduction It has been consistently observed that rapid human arm movements in both infants and adults often consist of several submovements, sometimes called ?movement units? [21]. The tangential hand velocity profiles of such movements show sequences of several bellshaped acceleration-deceleration phases, sometimes overlapping in the time domain and sometimes completely separate. Multiple movement units are observed mostly in infant reaching [5, 21] and in reaching movements by adult subjects in the face of difficult timeaccuracy requirements [15]. These data provide clues about how the nervous system efficiently produces fast and accurate movements in the presence of noise and significant feedback delay. Most modeling efforts concerned with movement units have addressed only the kinematic aspects of movement, e.g., [5, 12]. We show that multiple movement units might emerge as the result of a control policy that is optimal in the face of uncertainty and feedback delay. We use a simplified dynamic model of an arm and address rapid aimed arm movements. We use reinforcement learning as a tool to approximate the optimal policy in the presence of noise and feedback delay. An important motivation for this research is that stochasticity inherent in the motor control problem has a significant influence on the optimal control policy [9]. We are following the preliminary work of Zelevinsky [23] who showed that multiple movement units emerge from the stochasticity of the environment combined with a feedback delay. Whereas he restricted attention to a finite-state system to which he applied dynamic programming, our model has a continuous state space and we use reinforcement learning in a simulated realtime learning framework. 2 The model description The model we simulated is sketched in Figure 1. Two main parts of this model are the ?RL controller? (Reinforcement Learning controller) and the ?plant.? The controller here represents some functionality of the central nervous system dealing with the control of reaching movements. The plant represents a simplified arm together with spinal circuitry. The controller generates the control signal, , which influences how the state,  , of the plant changes over time. To simulate delayed feedback the state of the plant is made available to  the controller after a delay period , so at time  the controller can only observe  . To introduce stochasticity, we disturbed by adding noise to it, to produce a corrupted control . The controller learns to move the plant state as quickly as possible into a small region about a target state  . The reward structure block in Figure 1 provides a negative unit reward when the plant?s state is out of the target area of the state space, and it provides zero reward when the plant state is within the target area. The reinforcement learning controller tries to maximize the total cumulative reward for each movement. With the above mentioned reward structure, the faster the plant is driven into the target region, the less negative reward is accumulated during the movement. Thus this reward structure specifies the minimum time-to-goal criterion. r reward RL controller s u efferent copy u r reward target state s T structure state s delay noise ~ u target plant Figure 1: Sketch of the model used in our simulations. ?RL controller? stands for a Reinforcement Learning controller. 2.1 The plant To model arm dynamics together with the spinal reflex mechanisms we used a fractionalpower damping dynamic model [22]. The simplest model that captures the most critical dynamical features is a spring-mass system with a nonlinear damping: "!$#         Here,  is the position of the mass attached to the spring,  and   are respectively the velocity and the acceleration of the object,  is the mass of the object (the mass of the spring is assumed equal to zero),  is the damping coefficient,  is the stiffness coefficient, and is the control signal which determines the resting, or equilibrium, position. Later in this paper, we call activation, referring to the activation level of a muscle pair. The Table 1: Parameter values used in the simulations. description the basic simulation time step the feedback delay, initial position initial velocity target position target velocity target position radius  value 1 ms 200 ms 0 cm 0 cm/s 5 cm 5 cm 0.5 cm description threshold velocity radius standard deviation of the noise value function learning rate preferences learning rate discount factor, bootstrapping factor,    value 0.1 cm/s 1 cm 0.5 1 0.9 0.9    values for the mass, the damping coefficient, and the stiffness coefficient were taken from      ! Barto et al. [3]:  kg,  , . These values provide movement  trajectories qualitatively similar to those observed in human wrist movements [22]. The fractional-power damping in this model is motivated by both biological evidence [8, 14] and computational considerations. Controlling a system with such a concave damping function is an easier control problem than for a system with apparently simpler linear damping. Fractional-power damping creates a qualitatively novel dynamical feature called a stiction region, a region in the position space around the equilibrium position consisting of pseudo-stable states, where the velocity of the plant remains very close to zero. Such states are stable states for all practical purposes. For the parameter magnitudes used in our simulations, the stiction region is a region of radius 2.5 cm about the true equilibrium in the position space. Another essential feature of the neural signal transmission can be accounted for by using a cascade of low-pass temporal filters on the activation level [16]. We used a second-order low-pass filter with the time constant of 25 ms. 2.2 The reinforcement learning controller We used the version of the actor-critic algorithm described by Sutton and Barto [20]. A possible model of how an actor-critic architecture might be implemented in the nervous system was suggested by Barto [2] and Houk et al. [10]. We implemented the actor-critic algorithm for a continuous state space and a finite set of actions, i.e., activation level magnitudes evenly spaced every 1 cm between 0 cm and 10 cm. To represent functions defined over the continuous state space we used a CMAC representation [1] with 10 tilings, each tiling spans all three dimensions of the state space and has 10 tiles per dimension. The tilings have random offsets drawn from the uniform distribution. Learning is done in episodes. At the beginning of each episode the plant is at a fixed initial state, and the episode is complete when the plant reaches the target region of the state space. Table 1 shows the parameter values used in the simulations. Refer to ref. [20] for algorithm details. 2.3 Clocking the control signal For the controller to have sufficient information about the current state of the plant, the controller internal representation of the state should be augmented by a vector of all the actions selected during the last delay period. To keep the dimension of the state space at a feasible level, we restrict the set of available policies and make the controller select a new activation level, , in a clocked manner at time intervals equal to the delay period. One step of the reinforcement learning controller is performed once a delay period, which corresponds to many steps of the underlying plant simulation. To simulate a stochastic plant we added Gaussian noise to . A new Gaussian disturbance was drawn every time a new activation level was selected. Apart from the computational motivation, there is evidence of intermittent motor control by human subjects [13]. In our simulations we use an oversimplified special kind of intermittent control with a piecewise constant control signal whose magnitude changes at equal time intervals, but this is done for the sake of acceleration of the simulations and overall clarity. Intermittent control does not necessarily assume this particular kind of the control signal; the most important feature is that control segments are selected at particular points in time, and each control segment determines the control signal for an extended time interval. The time interval until selection of the next control segment can itself be one of the parameters [11]. 3 Results The model learned to move the mass quickly and accurately to the target in approximately 1,000 episodes. Figure 2 shows the corresponding learning curve. Figure 3 shows a typical movement accomplished by the controller after learning. The movement shown in Figure 3 has two acceleration-deceleration phases called movement units or submovements. 4000 3500 time per episode, ms 3000 2500 2000 1500 1000 500 0 0 100 200 300 400 500 episode # 600 700 800 900 1000 Figure 2: The learning curve averaged over 100 trials. The performance is measured in time-per-episode. Corrective submovements may occur before the plant reaches zero velocity. The controller generates this corrective submovement ?on the fly,? i.e., before the initial fast submovement is completed. Figure 4 shows a sample movement accomplished by the controller after learning where such overlapping submovements occur. This can be seen clearly in panel (b) of Figure 4 where the velocity profile of the movement is shown. Each of the submovements appears as a bell-shaped unit in the tangential velocity plot. Sometimes the controller accomplishes a movement with a single smooth submovement. A sample of such a movement is shown in Figure 5. 4 Discussion The model learns to produce movements that are fast and accurate in the presence of noise and delayed sensory feedback. Most of the movements consist of several submovements. The first submovement is always fast and covers most of the distance from the initial po- (d) 25 4 20 2 0 0 200 400 600 t, ms 800 1000 1200 (b) velocity, cm/s position, cm (a) 6 15 10 5 velocity, cm/s 30 0 20 10 ?5 0 ?10 0 200 400 600 t, ms 800 1000 1200 800 1000 1200 0 1 2 3 position, cm 4 5 6 (c) activation, cm 15 10 5 0 0 200 400 600 t, ms Figure 3: A sample movement accomplished by the controller after learning. Panels (a) and (b) show the position and velocity time course respectively. Panel (c) shows the activation time courses. The thin solid line shows the activation selected by the controller. The thick solid line shows the disturbed activation which is sent as the control signal to the plant. The dashed line shows the activation after the temporal filtering is applied. Panel (d) shows the phase trajectory of the movement. The thick bar at the lower-right corner is the target region. sition to the target. All of the subsequent submovements are much slower and cover much shorter segments in the position space. This feature stands in good agreement with the dual control model [12, 17], where the initial part of a movement is conducted in a ballistic manner, and the final part is conducted under closed-loop control. Some evidence for this kind of dual control strategy comes from experiments in which subjects were given visual feedback only during the initial stage of movement. Subjects did not show significant improvement under these conditions compared to trials in which they were deprived of visual feedback during the entire movement [4, 6]. In another set of experiments, proprioceptive feedback was altered by stimulations of muscle tendons. Movement accuracy decreased only when the stimulation was applied at the final stages of movement [18]. Note, however, that the dual control strategy though is not explicitly designed into our model, but naturally emerges from the existing constraints and conditions. The reinforcement learning controller is encouraged by the reward structure to accomplish each movement as quickly as possible. On the other hand, it faces high uncertainty in the plant behavior. In states with low velocities the information available to the controller determines the actual state of the plant quite accurately as opposed to states with high (a) (d) 20 4 15 2 0 0 200 400 600 800 1000 t, ms velocity, cm/s position, cm 6 10 5 (b) velocity, cm/s 20 0 15 10 ?5 5 0 0 200 400 600 800 1000 600 800 1000 0 1 2 3 position, cm 4 5 6 t, ms (c) activation, cm 15 10 5 0 0 200 400 t, ms Figure 4: A sample movement accomplished by the controller after learning with a well expressed predictive correction. velocities. If the controller were to adopt a policy in which it attempts to directly hit the target in one fast submovement, then very often it would miss the target and spend long additional time to accomplish the task. The optimal policy in this situation is to move the arm close to the target by one fast submovement and then apply a few slow submovements to accurately move arm into the target region. The model learns to produce control sequences consisting of pairs of high activation steps followed by low activation steps. This feature stands in good agreement with pulse-step models of motor control [7, 19]. Each of the pulse-step combinations produces a submovement characterized by a bell-shaped unit in the velocity profile. In biological motor control corrective submovements are observed very consistently, including both the overlapping and separate submovements. In the case of overlapping submovements, the corrective movement is called a predictive correction. Multiple submovements are observed mostly in infant reaching [5]. Adults perform routine everyday reaching movements ordinarily with a single smooth submovement, but in case of tight time constraints or accuracy requirements they revert to multiple submovements [15]. The suggested model sometimes accomplishes movements with a single smooth submovement (see Figure 5), but in most cases it produces multiple submovements much like an infant or an adult subject trying to move quickly and accurately. The suggested model is also consistent with theories of basal ganglia information processing for motor control [10]. Some of these theories suggest that dopamine neurons in the basal ganglia carry information similar to the secondary reinforcement (or temporal difference) in the actor-critic controller, i.e., information about how the expected perfor- (a) (d) 20 4 15 2 0 0 200 400 t, ms 600 800 velocity, cm/s position, cm 6 10 5 (b) velocity, cm/s 20 0 15 10 ?5 5 0 0 200 400 t, ms 600 800 600 800 0 1 2 3 position, cm 4 5 6 (c) activation, cm 15 10 5 0 0 200 400 t, ms Figure 5: A sample movement accomplished by the controller after learning with a single smooth submovement. mance (time-to-target) changes over time during a movement. A possible use of this kind of information is for initiating corrective submovements before the current movement is completed. This kind of behavior is exhibited by our model (Figure 4). Acknowledgments This work was supported by NIH Grant MH 48185?09. We thank Andrew H. Fagg and Michael T. Rosenstein for helpful comments. References [1] J. S. Albus. A new approach to manipulator control: the cerebellar model articulation controller (CMAC). Journal of Dynamics, Systems, Measurement and Control, 97:220?227, 1975. [2] A. G. Barto. Adaptive critics and the basal ganglia. In J. C. Houk, J. L. Davis, and D. G. Beiser, editors, Models of Information Processing in the Basal Ganglia, pages 215?232. MIT Press, Cambridge, MA, 1995. [3] A. G. Barto, A. H. Fagg, N. Sitkoff, and J. C. Houk. A cerebellar model of timing and prediction in the control of reaching. Neural Computation, 11:565?594, 1999. [4] D. Beaubaton and L. Hay. Contribution of visual information to feedforward and feedback processes in rapid pointing movements. Human Movement Science, 5:19?34, 1986. [5] N. E. Berthier. Learning to reach: a mathematical model. Developmental Psychology, 32:811? 832, 1996. [6] L. G. Carlton. Processing of visual feedback information for movement control. Journal of Experimental Psychology: Human Perception and Performance, 7:1019?1030, 1981. [7] C. Ghez. Contributions of central programs to rapid limb movement in the cat. In H. Asanuma and V. J. Wilson, editors, Integration in the Nervous System, pages 305?320. Igaku-Shoin, Tokyo, 1979. [8] C. C. A. M. Gielen and J. C. Houk. A model of the motor servo: incorporating nonlinear spindle receptor and muscle mechanical properties. Biological Cybernetics, 57:217?231, 1987. [9] C. M. Harris and D. M. Wolpert. Signal-dependent noise determines motor planning. Nature, 394:780?784, 1998. [10] J. C. Houk, J. L. Adams, and A. G. Barto. A model of how the basal ganglia generates and uses neural signals that predict reinforcement. In J. C. Houk, J. L. Davis, and D. G. Beiser, editors, Models of Information Processing in the Basal Ganglia, pages 249?270. MIT Press, Cambridge, MA, 1995. [11] M. Kositsky. Motor Learning and Skill Acquisition by Sequences of Elementary Actions. PhD thesis, The Weizmann Institute of Science, Israel, October 1998. [12] D. E. Meyer, S. Kornblum, R. A. Abrams, C. E. Wright, and J. E. K. Smith. Optimality in human motor performance: ideal control of rapid aimed movements. Psychological Review, 95(3):340?370, 1988. [13] R. C. Miall, D. J. Weir, and J. F. Stein. Intermittency in human manual tracking tasks. Journal of Motor Behavior, 25:53?63, 1993. [14] L. E. Miller. Reflex stiffness of the human wrist. Master?s thesis, Department of Physiology, Northwestern University, Evanston, IL, 1984. [15] K. E. Novak, L. E. Miller, and J. C. Houk. Kinematic properties of rapid hand movements in a knob turning task. Experimental Brain Research, 132:419?433, 2000. [16] L. D. Partridge. Integration in the central nervous system. In J. H. U. Brown and S. S. Gann, editors, Engineering Principles in physiology, pages 47?98. Academic Press, New York, 1973. [17] R. Plamondon and A. M. Alimi. Speed/accuracy trade-offs in target-directed movements. Behavioral and Brain Science, 20:279?349, 1997. [18] C. Redon, L. Hay, and J.-L. Velay. Proprioceptive control of goal directed movements in man studied by means of vibratory muscle tendon stimulation. Journal of Motor Behavior, 23:101? 108, 1991. [19] D. A. Robinson. Oculomotor control signals. In G. Lennerstrand and P. B. y Rita, editors, Basic Mechanisms of Ocular Mobility and Their Clinical Implications, pages 337?374. Pergamon Press, Oxford, 1975. [20] R. S. Sutton and A. G. Barto. Reinforcement Learning: An Introduction. MIT Press, Cambridge, MA, 1998. [21] C. von Hofsten. Structuring of early reaching movements: A longitudinal study. Journal of Motor Behavior, 23:280?292, 1991. [22] C. H. Wu, J. C. Houk, K. Y. Young, and L. E. Miller. Nonlinear damping of limb motion. In J. M. Winters and S. L.-Y. Woo, editors, Multiple Muscle Systems: Biomechanics and Movement Organization, pages 214?235. Springer-Verlag, New York, 1990. [23] L. Zelevinsky. Does time-optimal control of a stochastic system with sensory delay produce movement units? Master?s thesis, University of Massachusetts, Amherst, 1998.
2080 |@word trial:2 version:1 pulse:2 simulation:9 solid:2 carry:1 initial:8 uma:1 longitudinal:1 existing:1 current:2 activation:15 subsequent:1 berthier:1 motor:14 plot:1 designed:1 infant:4 selected:4 nervous:6 beginning:1 smith:1 provides:2 preference:1 simpler:1 mathematical:1 novak:1 asanuma:1 behavioral:1 lennerstrand:1 manner:2 introduce:1 expected:1 rapid:8 behavior:5 planning:1 brain:2 oversimplified:1 initiating:1 actual:1 underlying:1 panel:4 mass:6 israel:1 kg:1 cm:26 kind:5 bootstrapping:1 pseudo:1 temporal:3 every:2 concave:1 hit:1 evanston:1 control:39 unit:12 grant:1 appear:1 before:4 engineering:1 timing:1 sutton:2 receptor:1 oxford:1 approximately:1 might:3 studied:1 suggests:1 averaged:1 weizmann:1 directed:2 practical:1 acknowledgment:1 wrist:2 block:1 cmac:2 area:2 bell:3 cascade:1 physiology:2 suggest:1 close:3 selection:1 influence:2 disturbed:2 deterministic:1 attention:1 deceleration:3 target:21 controlling:1 programming:1 us:1 rita:1 agreement:2 velocity:20 observed:6 fly:1 capture:1 region:10 episode:7 hofsten:1 movement:55 trade:1 servo:1 mentioned:1 environment:1 developmental:1 reward:10 dynamic:6 tight:1 segment:4 predictive:3 creates:1 completely:1 po:1 mh:1 cat:1 corrective:6 revert:1 fast:8 whose:1 quite:1 spend:1 emergence:1 itself:1 final:2 sequence:4 loop:1 albus:1 description:3 everyday:1 bellshaped:1 requirement:2 transmission:1 produce:7 adam:1 object:2 andrew:2 measured:1 implemented:2 c:1 come:1 radius:3 thick:2 tokyo:1 functionality:1 filter:2 stochastic:2 human:9 rosenstein:1 preliminary:1 biological:3 elementary:1 correction:3 around:1 wright:1 houk:8 equilibrium:3 predict:1 circuitry:1 pointing:1 adopt:1 early:1 purpose:1 ballistic:1 tool:2 mit:3 clearly:1 offs:1 gaussian:2 always:1 reaching:7 barto:9 wilson:1 knob:1 structuring:1 improvement:1 consistently:2 helpful:1 dependent:1 accumulated:1 entire:1 sketched:1 overall:1 dual:3 special:1 integration:2 equal:3 once:1 shaped:3 encouraged:1 represents:2 thin:1 piecewise:1 inherent:1 tangential:3 few:2 winter:1 delayed:2 zelevinsky:2 phase:4 consisting:2 attempt:1 organization:1 kinematic:2 clocking:1 implication:1 accurate:2 shorter:1 mobility:1 damping:9 psychological:1 modeling:1 cover:2 deviation:1 uniform:1 delay:15 conducted:2 corrupted:1 accomplish:2 combined:1 referring:1 spindle:1 amherst:2 michael:2 together:2 quickly:4 ghez:1 thesis:3 central:3 von:1 opposed:1 tile:1 corner:1 coefficient:4 explicitly:1 later:1 try:1 performed:1 closed:1 apparently:1 contribution:2 il:1 accuracy:3 who:1 efficiently:2 miller:3 spaced:1 accurately:5 trajectory:2 drive:2 cybernetics:1 reach:3 manual:1 acquisition:1 ocular:1 naturally:1 efferent:1 massachusetts:2 fractional:2 emerges:1 routine:1 appears:1 done:2 though:1 stage:2 until:1 hand:4 sketch:1 nonlinear:3 overlapping:4 manipulator:1 brown:1 true:1 proprioceptive:2 during:5 davis:2 clocked:1 criterion:1 m:13 trying:1 complete:1 motion:1 consideration:1 novel:1 nih:1 stimulation:3 rl:3 spinal:2 attached:1 he:2 resting:1 significant:3 refer:1 measurement:1 cambridge:3 stochasticity:4 stable:2 actor:4 showed:1 driven:1 apart:1 hay:2 verlag:1 carlton:1 accomplished:5 muscle:5 seen:1 minimum:1 additional:1 accomplishes:2 maximize:1 period:4 signal:11 dashed:1 beiser:2 multiple:9 smooth:4 faster:1 characterized:1 academic:1 clinical:1 long:1 biomechanics:1 prediction:1 basic:2 controller:30 essentially:1 sition:1 dopamine:1 represent:1 sometimes:6 cerebellar:2 whereas:1 decreased:1 addressed:1 interval:4 exhibited:1 comment:1 subject:5 sent:1 call:1 presence:7 ideal:1 feedforward:1 concerned:1 psychology:2 architecture:1 restrict:1 motivated:1 effort:1 york:2 action:3 aimed:3 discount:1 stein:1 simplest:1 specifies:1 per:3 basal:6 threshold:1 drawn:2 clarity:1 fagg:2 uncertainty:2 master:2 wu:1 realtime:1 tendon:2 followed:1 occur:2 constraint:2 sake:1 generates:4 aspect:1 simulate:2 speed:1 span:1 spring:3 optimality:1 department:2 combination:1 deprived:1 restricted:1 taken:1 remains:1 mechanism:2 end:2 tiling:3 available:3 mance:1 stiffness:3 apply:2 observe:1 limb:2 slower:1 completed:3 psychophysical:1 move:5 pergamon:1 added:1 strategy:2 abrams:1 distance:1 separate:2 thank:1 simulated:2 evenly:1 difficult:1 mostly:2 october:1 negative:2 ordinarily:1 policy:11 perform:1 observation:1 neuron:1 finite:2 situation:2 extended:1 intermittency:1 intermittent:3 pair:2 mechanical:1 learned:1 robinson:1 address:2 adult:4 suggested:3 bar:1 dynamical:2 perception:1 articulation:1 oculomotor:1 program:1 including:1 power:2 critical:2 disturbance:1 turning:1 arm:12 altered:1 woo:1 review:1 plant:21 northwestern:1 filtering:1 sufficient:1 consistent:1 principle:1 editor:6 critic:5 course:2 accounted:1 supported:1 last:1 copy:1 institute:1 face:3 emerge:3 feedback:16 dimension:3 curve:2 stand:3 cumulative:1 sensory:2 made:1 reinforcement:12 clue:1 simplified:4 qualitatively:2 adaptive:1 miall:1 approximate:2 skill:1 keep:1 dealing:1 assumed:1 continuous:3 table:2 nature:1 necessarily:1 domain:1 did:1 main:1 motivation:2 noise:12 profile:4 ref:1 augmented:1 slow:2 position:16 meyer:1 learns:3 young:1 offset:1 evidence:3 consist:2 essential:1 incorporating:1 adding:1 phd:1 magnitude:3 easier:1 wolpert:1 gielen:1 ganglion:6 visual:4 expressed:1 tracking:1 reflex:2 springer:1 corresponds:1 determines:4 harris:1 ma:4 goal:2 acceleration:5 weir:1 man:1 feasible:1 change:3 plamondon:1 vibratory:1 typical:1 miss:1 called:5 total:1 pas:2 secondary:1 experimental:2 select:1 perfor:1 internal:1
1,186
2,081
Fragment completion in humans and machines David Jacobs NEC Research Institute 4 Independence Way, Princeton, NJ 08540 dwj@research.nj.nec.com Archisman Rudra CS Department at NYU 251 Mercer St., New York, NY 10012 archi@cs.nyu.edu Bas Rokers Psychology Department at UCLA PO Box 951563, Los Angeles, CA 90095 rokers@psych.ucla.edu Zili Liu Psychology Department at UCLA PO Box 951563, Los Angeles CA 90095 zili@psych.ucla.edu Abstract Partial information can trigger a complete memory. At the same time, human memory is not perfect. A cue can contain enough information to specify an item in memory, but fail to trigger that item. In the context of word memory, we present experiments that demonstrate some basic patterns in human memory errors. We use cues that consist of word fragments. We show that short and long cues are completed more accurately than medium length ones and study some of the factors that lead to this behavior. We then present a novel computational model that shows some of the flexibility and patterns of errors that occur in human memory. This model iterates between bottom-up and top-down computations. These are tied together using a Markov model of words that allows memory to be accessed with a simple feature set, and enables a bottom-up process to compute a probability distribution of possible completions of word fragments, in a manner similar to models of visual perceptual completion. 1 Introduction This paper addresses the problem of retrieving items in memory from partial information. Human memory is remarkable for its flexibility in handling a wide range of possible retrieval cues. It is also very accurate, but not perfect; some cues are more easily used than others. We hypothesize that memory errors occur in part because a trade-off exists between memory accuracy and the complexity of neural hardware needed to perform complicated memory tasks. If this is true, we can gain insight into mechanisms of human memory by studying the patterns of errors humans make, and we can model human memory with systems that produce similar patterns as a result of constraints on computational resources. We experiment with word memory questions of the sort that arise in a game called superghost. Subjects are presented with questions of a form: ?*p*l*c*?. They must find a valid English word that matches this query, by replacing each ?*? with zero or more letters. So for this example, ?place?, ?application?, and ?palace? would all be valid answers. In ef- fect, the subject is given a set of letters and must think of a word that contains all of those letters, in that order, with other letters added as needed. Most of the psychological literature on word completion involves the effects of priming certain responses with recent experience (Shacter and Tulving[18]). However, priming is only able to account for about five percent of the variance in a typical fragment completion task (Olofsson and Nyberg[13], Hintzman and Hartry[6]). We describe experiments that show that the difficulty of a query depends on what we call its redundancy. This measures the extent to which all the letters in the query are needed to find a valid answer. We show that when we control for the redundancy of queries, we find that the difficulty of answering questions increases with their length; queries with many letters tend to be easy only because they tend to be highly redundant. We then describe a model that mimics these and other properties of human memory. Our model is based on the idea that a large memory system can gain efficiency by keeping the comparison between input and items in memory as simple as possible. All comparisons use a small, fixed set of features. To flexibly handle a range of queries, we add a bottomup process that computes the probability that each feature is present in the answer, given the input and a generic, Markov model of words. So the complexity of the bottom-up computation does not grow with the number of items in memory. Finally, the system is allowed to iterate between this bottom up and a top down process, so that a new generic model of words is constructed based on a current probability distribution over all words in memory, and this new model is combined with the input to update the probability that each feature is present in the answer. Previous psychological research has compared performance of word-stem and wordfragment completion. In the former a number of letters (i.e. a fragment) is given beginning with the first letter(s) of the word. In the latter, the string of letters given may begin at any point in the word, and adjacent letters in the fragment do not need, but may, be adjacent in the completed word. For example, for stem completion the fragment ?str? may be completed into ?string?, but for fragment completion also into ?satire?. Performance for wordfragment completion is lower than word-stem completion (Olofsson and Nyberg[12]). In addition words, for which the ending fragment is given, show performance closer to wordstem completion than to word-fragment completion (Olofsson and Nyberg[13]). Seidenberg[17] proposed a model based on tri-grams. Srinivas et al.[21] indicate that assuming orthographic encoding is in most cases sufficient to describe word completion performance in humans. Orthographic Markov models of words have often been used computationally, as, for example, in Shannon?s[19] famous work. Following this work, our model is also orthographic. We find that a bigram rather than a trigram representation is sufficient, and leads to a simpler model. Contradicting evidence exists for the influence of fragment length on word completion. Oloffsson and Nyberg [12] failed to find a difference between two and three letter fragments on words of length of five to eight letters. However this might have been due to the fact that in their task, each fragment has a unique completion. Many recurrent neural networks have been proposed as models of associative memory (Anderson[1] contains a review). Perhaps most relevant to our work are models that use an input query to activate items from a complete dictionary in memory, and then use these items to alter the activations of the input. For example, in the Interactive Activation model of Rumelhart and McClelland[16], the presence of letters activates words, which boost the activity of the letters they contain. In Adaptive Resonance models (Carpenter and Grossberg[3]) activated memory items are compared to the input query and de-activated if they do not match. Also similar in spirit to our approach is the bidirectional model of Kosko[10] (for more recent work see, eg., Sommer and Palm[20]). Other models iteratively combine top-down and bottom-up information (eg., Hinton et al.[5], Rao and Ballard[14]), although these are not used as part of a memory system with complete items stored in memory. Our model differs from all of these in using a Markov model as an intermediate layer between the input and the dictionary. This allows the model to answer superghost queries, and leads to different computational mechanisms that we will detail. We find that superghost queries seem more natural to people than associative memory word problems (compare the superghost query ?think of a word with an a? to the associative memory query ?think of a word whose seventh letter is an a?). However, it is not clear how to extend most models of associative memory to handle superghost problems. Our use of features is more related to feedforward neural nets, and especially the ?information bottleneck? approach of Tishby, Pereira and Bialek[22] (see also Baum, et al.[2]). Our work differs from feedforward methods in that our method is iterative, and uses features symmetrically to relate the memory to input in both directions. Our approach is also related to work on visual object recognition that combines perceptual organization and top-down knowledge (see Ullman[23]). Our model is inspired by Mumford?s[11] and Williams and Jacobs?[24] use of Markov models of contours for bottom-up perceptual completion. Especially relevant to our work is that of Grimes and Mozer[4]. Simultaneous with our work ([8]) they use a bigram model to solve anagram problems, in which letters are unscrambled to match words in a dictionary. They also use a Markov model to find letter orderings that conform with the statistics of English spelling. Their model is quite different in how this is done, due to the different nature of the anagram problem. They view anagram solving as a mix of low-level processing and higher level cognitive processes, while it is our goal to focus just on lower level memory. 2 Experiments with Human Subjects In our experiments, fragments and matching words were drawn from a large standard corpus of English text. The frequency of a word is the number of times it appears in this corpus. The frequency of a fragment is the sum of the frequency of all words that the fragment matches. We used fragments of length two to eight, discarding any fragments with frequency lower than one thousand. Fragments selected for an experiment were presented in random order. In our first experiment we systematically varied the length of the fragments, but otherwise selected them from a uniform, random distribution. Consequently, shorter fragments tended to match more words, with greater total frequency. In the second experiment, fragments were selected so that a uniform distribution of frequencies was ensured over all fragment lengths. For example, we used length two fragments that matched unusually few words. As a result the average frequency in experiment two is also much lower than in experiment one. A fragment was presented on a computer screen with spaces interspersed, indicating the possibility of letter insertion. The subject was required to enter a word that would fit the fragment. A subject was given 10 seconds to produce a completion, with the possibility to give up. For each session 50 fragments were presented, with a similar number of fragments of each length. Reaction times were recorded by measuring the time elapsed between the fragment first appearing on screen and the subject typing the first character of a matching word. Words that did not match the fragment or did not exist in the corpus were marked as not completed. Each experiment was completed by thirty-one subjects. The subjects were undergraduate students at Rutgers University, participating in the experiment for partial credit. Total time 1 0.8 Fraction Completed Fraction Completed 1 0.8 0.6 0.6 0.4 0.4 0.2 0.2 0 1 2 3 4 5 6 7 8 0 1 9 Fragment Length 2 3 4 5 6 Fragment Length 7 8 9 1 Fraction Completed 0.8 R1 0.6 R0 R4 0.4 R3 R2 0.2 0 1 2 3 4 5 6 7 8 9 Fragment Length Figure 1: Fragment completion as a function of fragment length for randomly chosen cues (top-left) and cues of equal frequency (top-right). On the bottom, the equal frequency cues are divided into five groups, from least redundancy (R0) to most (R5) . spent on the task varied from 15 minutes to close to one hour. Results For each graph we plot the number of fragments completed divided by the number of    , where is the fragments presented (Figure 1). Error bars are calculated as percent correct in the sample, and is the number of trials. This assumes that all decisions  are independent and correct with probability ; more precise results can be obtained by accounting for between-subject variance, but roughly the same results hold. For random, uniformly chosen fragments, there is a U-shaped dependence of performance on length. Controlling for frequency reduces performance because on average lower frequency fragments are selected. The U-shaped curve is flattened, but persists; hence Ushaped performance is not just due to frequency Finally, we divide the fragments from the two experiments into five groups, according to their redundancy. This is a rough measure of how important each letter is in finding a correct answer to the overall question. It is the probability that if we randomly delete a letter from the fragment and find a matching word, that this word will match the full fragment. Specifically, let  denote the frequency of a query fragment of length  (total frequency of words that match it). Let  denote the frequency of the fragment that results when we  delete the  ?th letter from the query (note,    ). Then redundancy is:   !"# . In all cases where there is a significant difference, greater redundancy leads to better performance. In almost all cases, when we control for redundancy performance decreases with length. We will discuss the implications of these experiments after describing corresponding experiments with our model. 3 Using Markov Models for Word Retrieval We now describe a model of word memory in which matching between the query and memory is mediated by a simple set of features. Specifically, we use bigrams (adjacent pairs of letters) as our feature set. We denote the beginning and end of a word using the symbols ?0? and ?1?, respectively, so that bigram probabilities also indicate how often individual letters begin or end a word. Bottom up processing of a cue is done using this as a Markov model of words. Then bigram probabilities are used to trigger words in memory that might match the query. Our algorithm consists of three steps. First, we compute a prior distribution on how likely each word in memory is to match our query. In our simulations, we just use a uniform distribution. However, this distribution could reflect the frequency with which each word occurs in English. It could also be used to capture priming phenomena; for example, if a word has been recently seen, its prior probability could increase, making it more likely that the model would retrieve this word. Then, using these we compute a probability that each bigram will appear if we randomly select a bigram from a word selected according to our prior distribution. Second, we use these bigram probabilities as a Markov model, and compute the expected number of times each bigram will occur in the answer, conditioned on the query. That is, as a generic model of words we assume that each letter in the word depends on the adjacent letters, but is conditionally independent of all others. This conditional independence allows us to decompose our problem into a set of small, independent problems. For example, consider the query ?*p*l*c*?. Implicitly, each query begins with ?0? and ends with ?1?, so the expected number of times any bigram will appear in the completed word is the sum of the number of times it appears in the completions of the fragments: ?0*p?, ?p*l?, ?l*c?, and ?c*1?. To compute this, we assume a prior distribution on the number of letters that will replace a ?*? in the completed word. We use an exponential model, setting the probability of letters   to be  (in practice we truncate at 5 and normalize the probabilities). A similar model is used in the perceptual completion of contours ([11, 24]). Using these priors, it becomes straightforward to compute a probability distribution on the bigrams that will appear in the  bigrams, completed cue. For a fixed , we structure this problem as a belief net with  and each bigram depending on only its neighbors. The conditional probability of each bigram given its neighbor comes from the Markov model, and we can solve the problem with belief propagation. Beginning the third step of the algorithm, we know the expected number of times that each bigram appears in the completed cue. Each bigram then votes for all words containing that bigram. The weight of this vote is the expected number of times each bigram appears in the completed cue, divided by the prior probability of each bigram, computed in step 1. We combine these votes multiplicatively. We update the prior for each word as the product of these votes with the previous probability. We can view this an approximate computation of the probability of each word being the correct answer, based on the likelihood that a bigram appears in the completed cue, and our prior on each word being correct. After the third step, we once again have a probability that each word is correct, and can iterate, using this probability to initialize step one. After a small number of iterations, we terminate the algorithm and select the most probable word as our answer. Empirically, we find that the answer the algorithm produces often changes in the first one or two iterations, and then generally remains the same. The answer may or may not actually match the input cue, and by this we judge whether it is correct or incorrect. We can view this algorithm as an approximate computation of the probability that each 1 0.9 0.9 Fraction Completed Fraction Completed 1 0.8 0.8 0.7 0.7 0.6 0.6 0.5 0.5 0.4 1 2 3 4 5 6 Fragment Length 7 8 0.4 1 9 2 3 4 5 6 7 8 9 Fragment Length 1 R4 R2 0.8 R1 Fraction Completed 0.9 R3 0.7 0.6 R0 0.5 0.4 1 2 3 4 5 6 7 8 9 Fragment Length Figure 2: Performance as a function of cue length, for cues of frequency between 4 and 22 (top-left) and between 1 and 3 (top-right). On the bottom, we divide the first set of cues into five groups ranging from the least redundant (R0) to the most (R4). word matches the cue, where the main approximation comes from using a small set of features to bring the cue into contact with items in memory. Denote the number of features ), the number of features in each word by by (with a bigram representation, (ie., the word length plus one), the number of words by , and the maximum number of blanks replacing a ?*? by . Then steps one and three require O(mw) computation, and step two requires O(Fn) computation. In a neural network, the primary requirement would be bidirectional connections between each feature (bigram) and each item in memory. Therefore, computational simplicity is gained by using a small feature set, at the cost of some approximation in the computation.  Experiments We have run experiments to compare the performance of this model to that of human subjects. For simplicity, we used a memory of 6,040 words, each with eight characters. First, we simulated the conditions described in Olofsson and Nyberg[12] comparing word stem and word fragment completion. To match their experiments, we used a modified algorithm that handled cues in which the number of missing letters can be specified. We used cues that specified the first three letters of a word, the last three letters, or three letters scattered throughout the word. The algorithm achieved accuracy of 95% in the first case, 87% in the second, and 80% in the third. This qualitatively matches the results for human subjects. Note that our algorithm treats the beginning and end of words symmetrically. Therefore, the fact that it performs better when the first letters of the word are given than when the last are given is due to regularities in English spelling, and is not built into the algorithm. Next we simulated conditions comparable to our own experiments on human subjects, using superghost cues. First we selected cues of varying length that match between four and twenty-two words in the dictionary. Figure 2-top-left shows the percentage of queries the algorithm correctly answered, for cues of lengths two to seven. This figure shows a U-shaped performance curve qualitatively similar to that displayed by human subjects. We also ran these experiments using cues that matched one to three words (Figure 2-topright). These very low frequency cues did not display this U-shaped behavior. The algorithm performs differently on fragments with very low frequency because in our corpus the shorter of these cues had especially low redundancy and the longer fragments had especially high redundancy, in comparison to fragments with frequencies between 4 and 22. Next (Figure 2-bottom) we divided the cues into five groups of equal size, according to their redundancy. We can see that performance increases with redundancy and decreases with cue length. Discussion Our experiments indicate two main effects in human word memory that our model also shares. First, performance improves with the redundancy of cues. Second, when we control for this, performance drops with cue length. Since redundancy tends to increase with cue length, this creates two conflicting tendencies that result in a U-shaped memory curve. We conjecture that these factors may be present in many memory tasks, leading to U-shaped memory curves in a number of domains. In our model, the fact that performance drops with cue length is a result of our use of a simple feature set to mediate matching the cue to words in memory. This means that not all the information present in the cue is conveyed to items in memory. When the length of a cue increases, but its redundancy remains low, all the information in the cue remains important in getting a correct answer, but the amount of information in the cue increases, making it harder to capture it all with a limited feature set. This can account for the performance of our model; similar mechanisms may account for human performance as well. On the other hand, the extent to which redundancy grows with cue length is really a product of the specific words in memory and the cues chosen. Therefore, the exact shape of the performance curve will also depend on these factors. This may partly explain some of the quantitative differences between our model and human performance. Finally, we also point out that our measure of redundancy is rather crude. In particular, it tends to saturate at very high or very low levels. So, for example, if we add a letter to a cue that is already highly redundant, the new letter may not be needed to find a correct answer, but that is not reflected by much of an increase in the cue?s redundancy. 4 Conclusions We have proposed superghost queries as a domain for experimenting with word memory, because it seems a natural task to people, and requires models that can flexibly handle somewhat complicated questions. We have shown that in human subjects, performance on superghost improves with the redundancy of a query, and otherwise tends to decrease with word length. Together, these effects results in a U-shaped performance curve. We have proposed a computational model that uses a simple, generic model of words to map a superghost query onto a simple feature set of bigrams. This means that somewhat complicated questions can be answered while keeping comparisons between the fragments and words in memory very simple. Our model displays the two main trends we have found in human memory. It also does better at word stem completion than word fragment completion, which agrees with previous work on human memory. Future work will investigate the modification of our model to account for priming effects in memory. References [1] J. Anderson. An Introduction to Neural Networks, MIT Press, Cambridge MA. 1995. [2] E. Baum, J. Moody and F. Wilczek. ?Internal Representations for Associative Memory,? Biological Cybernetics, 59:217-228, 1988. [3] G. Carpenter, and S. Grossberg. ?ART 2: Self-Organization of Stable Category Recognition Codes for Analog Input Patterns,? Applied Optics, 26:4919-4930, 1987. [4] D. Grimes and M. Mozer. ?The interplay of symbolic and subsymbolic processes in anagram problem solving,? NIPS, 2001. [5] G. Hinton, P. Dayan, B. Frey, and R. Neal. ?The ?Wake-Sleep? Algorithm for Unsupervised Neural Networks,? Science, 268:1158-1161, 1995. [6] D.L. Hintzman and A.L. Hartry. Item effects in recognition and fragment completion: Contingency relations vary for different sets of words. JEP: Learning, Memory and Cognition, 17: 341-345, 1990. [7] J. Hopfield. ?Neural networks and Physical Systems with Emergent Collective Computational Abilities.? Proc. of the Nat. Acad. of Science, 79:2554-2558, 1982. [8] D. Jacobs and A. Rudra. ?An Iterative Projection Model of Memory,? NEC Research Institute Technical Report, 2000. [9] G.V. Jones. Fragment and schema models for recall. Memory and Cognition, 12(3):250-63, 1984. [10] B. Kosko. ?Adaptive Bidirectional Associative Memory?, Applied Optics, 26(23):4947-60, 1987. [11] D. Mumford. ?Elastica and Computer Vision.? C. Bajaj (Ed), Algebraic Geometry and its Applications New York: Springer-Verlag. 1994. [12] U. Olofsson and L. Nyberg. Swedish norms for completion of word stems and unique word fragments. Scandinavian Journal of Psychology, 33(2):108-16, 1992. [13] U. Olofsson and L. Nyberg. Determinants of word fragment completion. Scandinavian Journal of Psychology, 36(1):59-64, 1995. [14] R. Rao and D. Ballard. ?Dynamic Model of Visual Recognition Predicts Neural Response Properties in the Visual Cortex,? Neural Computation, 9(4):721-763, 1997. [15] R.H. Ross and G.H. Bower. Comparisons of models of associative recall. Memory and Cognition, 9(1):1-16, 1981. [16] D. Rumelhart and J. McClelland. ?An interactive activation model of context effects in letter perception: part 2. The contextual enhancement effect and some tests and extensions of the model?, Psychological Review, 89:60-94, 1982. [17] M.S. Seidenberg. Sublexical structures in visual word recognition: Access units or orthographic redundancy? In M. Coltheart (Ed.), Attention and performance XII, 245-263. Hillsdale, NJ: Erlbaum. 1987. [18] D.L. Shacter and E. Tulving. Memory systems. Cambridge, MA: MIT Press. 1994. [19] C. Shannon. ?Prediction and Entropy of Printed English,? Bell Systems Technical Journal, 30:50-64, 1951. [20] Sommer, F., and Palm, G., 1997, NIPS:676-681. [21] K. Srinivas, H.L. Roediger 3d and S. Rajaram. The role of syllabic and orthographic properties of letter cues in solving word fragments. Memory and Cognition, 20(3):219-30, 1992. [22] N. Tishby, F. Pereira and W. Bialek. ?The Information Bottleneck Method,? 37th Allerton Conference on Communication, Control, and Computing. 1999. [23] S. Ullman. High-level Vision, MIT Press, Cambridge, MA. 1996. [24] L. Williams & D. Jacobs. ?Stochastic Completion Fields: A Neural Model of Illusory Contour Shape and Salience?. Neural Computation, 9:837?858, 1997. Acknowledgements The authors would like to thank Nancy Johal for her assistance in conducting the psychological experiments presented in this paper.
2081 |@word trial:1 determinant:1 bigram:23 seems:1 norm:1 simulation:1 jacob:4 accounting:1 harder:1 liu:1 contains:2 fragment:59 reaction:1 current:1 com:1 blank:1 comparing:1 contextual:1 activation:3 must:2 fn:1 shape:2 enables:1 hypothesize:1 plot:1 drop:2 update:2 cue:43 selected:6 item:13 beginning:4 short:1 iterates:1 allerton:1 simpler:1 accessed:1 five:6 constructed:1 retrieving:1 incorrect:1 consists:1 shacter:2 combine:3 manner:1 anagram:4 expected:4 roughly:1 behavior:2 inspired:1 str:1 becomes:1 begin:3 matched:2 medium:1 what:1 psych:2 string:2 finding:1 nj:3 quantitative:1 unusually:1 interactive:2 coltheart:1 ensured:1 control:4 unit:1 appear:3 persists:1 frey:1 treat:1 tends:3 acad:1 encoding:1 might:2 plus:1 r4:3 limited:1 range:2 grossberg:2 unique:2 thirty:1 practice:1 orthographic:5 differs:2 jep:1 bell:1 printed:1 matching:5 projection:1 word:87 symbolic:1 onto:1 close:1 context:2 influence:1 map:1 missing:1 baum:2 williams:2 straightforward:1 flexibly:2 attention:1 simplicity:2 insight:1 retrieve:1 handle:3 controlling:1 trigger:3 exact:1 us:2 trend:1 rumelhart:2 recognition:5 predicts:1 bottom:10 role:1 capture:2 thousand:1 ordering:1 trade:1 decrease:3 ran:1 mozer:2 complexity:2 insertion:1 dynamic:1 depend:1 solving:3 creates:1 efficiency:1 po:2 easily:1 hopfield:1 differently:1 emergent:1 describe:4 activate:1 query:24 olofsson:6 whose:1 quite:1 solve:2 nyberg:7 otherwise:2 ability:1 statistic:1 think:3 associative:7 interplay:1 net:2 product:2 elastica:1 relevant:2 flexibility:2 participating:1 normalize:1 getting:1 los:2 regularity:1 requirement:1 r1:2 enhancement:1 produce:3 perfect:2 object:1 spent:1 depending:1 recurrent:1 completion:27 c:2 involves:1 indicate:3 come:2 judge:1 direction:1 correct:9 stochastic:1 human:21 hillsdale:1 require:1 really:1 decompose:1 probable:1 biological:1 extension:1 hold:1 credit:1 cognition:4 trigram:1 dictionary:4 vary:1 proc:1 ross:1 agrees:1 rough:1 mit:3 activates:1 modified:1 rather:2 varying:1 focus:1 likelihood:1 experimenting:1 dayan:1 her:1 relation:1 overall:1 resonance:1 art:1 initialize:1 equal:3 once:1 field:1 shaped:7 r5:1 jones:1 unsupervised:1 alter:1 mimic:1 future:1 others:2 rokers:2 report:1 few:1 palace:1 randomly:3 individual:1 geometry:1 organization:2 highly:2 possibility:2 investigate:1 grime:2 activated:2 implication:1 accurate:1 rudra:2 closer:1 partial:3 experience:1 shorter:2 divide:2 delete:2 psychological:4 rao:2 measuring:1 cost:1 uniform:3 seventh:1 erlbaum:1 tishby:2 stored:1 answer:13 combined:1 st:1 ie:1 off:1 together:2 moody:1 fect:1 reflect:1 recorded:1 again:1 containing:1 cognitive:1 leading:1 ullman:2 account:4 de:1 student:1 depends:2 view:3 schema:1 sort:1 complicated:3 accuracy:2 variance:2 conducting:1 rajaram:1 famous:1 accurately:1 cybernetics:1 simultaneous:1 explain:1 tended:1 ed:2 frequency:20 gain:2 illusory:1 recall:2 knowledge:1 nancy:1 improves:2 actually:1 appears:5 bidirectional:3 higher:1 reflected:1 specify:1 response:2 swedish:1 done:2 box:2 anderson:2 just:3 hand:1 replacing:2 wilczek:1 propagation:1 perhaps:1 grows:1 effect:7 contain:2 true:1 former:1 hence:1 iteratively:1 neal:1 eg:2 conditionally:1 adjacent:4 assistance:1 game:1 self:1 complete:3 demonstrate:1 performs:2 bring:1 percent:2 ranging:1 novel:1 ef:1 recently:1 physical:1 empirically:1 interspersed:1 extend:1 analog:1 significant:1 cambridge:3 enter:1 session:1 had:2 stable:1 scandinavian:2 longer:1 cortex:1 access:1 add:2 own:1 recent:2 certain:1 verlag:1 seen:1 greater:2 somewhat:2 r0:4 redundant:3 full:1 mix:1 reduces:1 stem:6 technical:2 match:15 long:1 retrieval:2 divided:4 prediction:1 basic:1 vision:2 rutgers:1 iteration:2 achieved:1 addition:1 wake:1 grow:1 tri:1 subject:14 tend:2 spirit:1 seem:1 call:1 mw:1 presence:1 symmetrically:2 intermediate:1 feedforward:2 enough:1 easy:1 iterate:2 independence:2 fit:1 psychology:4 zili:2 idea:1 angeles:2 bottleneck:2 whether:1 handled:1 algebraic:1 york:2 generally:1 clear:1 amount:1 hardware:1 mcclelland:2 category:1 exist:1 percentage:1 correctly:1 conform:1 xii:1 group:4 redundancy:19 four:1 drawn:1 graph:1 fraction:6 sum:2 run:1 letter:36 place:1 almost:1 throughout:1 decision:1 comparable:1 layer:1 display:2 sleep:1 activity:1 occur:3 optic:2 constraint:1 archi:1 ucla:4 answered:2 conjecture:1 department:3 palm:2 according:3 truncate:1 character:2 making:2 modification:1 satire:1 computationally:1 resource:1 remains:3 discus:1 r3:2 fail:1 mechanism:3 needed:4 describing:1 tulving:2 know:1 end:4 studying:1 eight:3 generic:4 appearing:1 top:9 assumes:1 sommer:2 completed:18 especially:4 contact:1 question:6 added:1 occurs:1 mumford:2 already:1 primary:1 dependence:1 spelling:2 bialek:2 thank:1 simulated:2 seven:1 dwj:1 extent:2 assuming:1 length:30 code:1 multiplicatively:1 relate:1 ba:1 collective:1 twenty:1 perform:1 kosko:2 markov:10 displayed:1 hinton:2 communication:1 precise:1 varied:2 david:1 pair:1 required:1 specified:2 connection:1 elapsed:1 conflicting:1 boost:1 hour:1 nip:2 address:1 able:1 bar:1 pattern:5 perception:1 built:1 memory:57 belief:2 difficulty:2 natural:2 typing:1 mediated:1 text:1 review:2 literature:1 prior:8 acknowledgement:1 remarkable:1 syllabic:1 contingency:1 conveyed:1 sufficient:2 mercer:1 systematically:1 share:1 last:2 keeping:2 english:6 salience:1 srinivas:2 institute:2 wide:1 neighbor:2 curve:6 calculated:1 valid:3 ending:1 gram:1 computes:1 contour:3 author:1 qualitatively:2 adaptive:2 approximate:2 implicitly:1 corpus:4 bottomup:1 iterative:2 seidenberg:2 ballard:2 nature:1 terminate:1 ca:2 roediger:1 priming:4 domain:2 did:3 main:3 arise:1 mediate:1 bajaj:1 contradicting:1 allowed:1 carpenter:2 screen:2 scattered:1 ny:1 pereira:2 exponential:1 crude:1 answering:1 tied:1 perceptual:4 third:3 bower:1 subsymbolic:1 down:4 minute:1 saturate:1 discarding:1 specific:1 symbol:1 nyu:2 r2:2 evidence:1 consist:1 exists:2 undergraduate:1 flattened:1 gained:1 nec:3 nat:1 conditioned:1 entropy:1 likely:2 visual:5 failed:1 springer:1 ma:3 hintzman:2 conditional:2 goal:1 marked:1 consequently:1 replace:1 change:1 typical:1 specifically:2 uniformly:1 called:1 total:3 partly:1 tendency:1 shannon:2 vote:4 indicating:1 select:2 internal:1 people:2 latter:1 princeton:1 phenomenon:1 handling:1
1,187
2,082
Global Coordination of Local Linear Models   Sam Roweis , Lawrence K. Saul , and Geoffrey E. Hinton Department of Computer Science, University of Toronto Department of Computer and Information Science, University of Pennsylvania Abstract High dimensional data that lies on or near a low dimensional manifold can be described by a collection of local linear models. Such a description, however, does not provide a global parameterization of the manifold?arguably an important goal of unsupervised learning. In this paper, we show how to learn a collection of local linear models that solves this more difficult problem. Our local linear models are represented by a mixture of factor analyzers, and the ?global coordination? of these models is achieved by adding a regularizing term to the standard maximum likelihood objective function. The regularizer breaks a degeneracy in the mixture model?s parameter space, favoring models whose internal coordinate systems are aligned in a consistent way. As a result, the internal coordinates change smoothly and continuously as one traverses a connected path on the manifold?even when the path crosses the domains of many different local models. The regularizer takes the form of a Kullback-Leibler divergence and illustrates an unexpected application of variational methods: not to perform approximate inference in intractable probabilistic models, but to learn more useful internal representations in tractable ones. 1 Manifold Learning Consider an ensemble of images, each of which contains a face against a neutral background. Each image can be represented by a point in the high dimensional vector space of pixel intensities. This representation, however, does not exploit the strong correlations between pixels of the same image, nor does it support many useful operations for reasoning about faces. If, for example, we select two images with faces in widely different locations and then average their pixel intensities, we do not obtain an image of a face at their average location. Images of faces lie on or near a low-dimensional, curved manifold, and we can represent them more usefully by the coordinates on this manifold than by pixel intensities. Using these ?intrinsic coordinates?, the average of two faces is another face with the average of their locations, poses and expressions. To analyze and manipulate faces, it is helpful to imagine a ?magic black box? with levers or dials corresponding to the intrinsic coordinates on this manifold. Given a setting of the levers and dials, the box generates an image of a face. Given an image of a face, the box deduces the appropriate setting of the levers and dials. In this paper, we describe a fairly general way to construct such a box automatically from an ensemble of high-dimensional vectors. We assume only that there exists an underlying manifold of low dimensionality and that the relationship between the raw data and the manifold coordinates is locally linear and smoothly varying. Thus our method applies not only to images of faces, but also to many other forms of highly distributed perceptual and scientific data (e.g., spectrograms of speech, robotic sensors, gene expression arrays, document collections). 2 Local Linear Models The global structure of perceptual manifolds (such as images of faces) tends to be highly nonlinear. Fortunately, despite their complicated global structure, we can usually characterize these manifolds as locally linear. Thus, to a good approximation, they can be represented by collections of simpler models, each of which describes a locally linear neighborhood[3, 6, 8]. For unsupervised learning tasks, a probabilistic model that nicely captures this intuition is a mixture of factor analyzers (MFA)[5]. The model is used to describe high dimensional data that lies on or near a lower dimensional manifold. MFAs parameterize a joint distribution over observed and hidden variables: (1) where the observed variable, , represents the high dimensional data; the discrete hidden variables, , indexes different neighborhoods on the manifold; and the continuous hidden variables, , represent low dimensional local coordinates. The model assumes that data is sampled from different neighborhoods on the manifold with prior probabilities , and that within each neighborhood, the data?s local coordinates are normally distributed1 as:          ! "$# % &')()(*(+-, . /0!1 243   256&87.:9 1;< =?>A@BDC &% *E F ( (2) G H I    J &7 I 8 9K ;< =?>A@LBDC &% M  C G C H :  N E I 9 K M  C G C H :  N F ( (3) O , is obtained by summing/integrating out the model?s The marginal data distribution, Finally, the model assumes that the data?s high and low dimensional coordinates are related by linear processes parameterized by centers , loading matrices and noise levels : discrete and continuous latent variables. The result is a mixture of Gaussian distributions with parameterized covariance matrices of the form: 2QP R3 - &872 H H "E S I ) 9 K ;< =)>T@ B C L&% M  C G N E  H H E S I  9 K M  C G N:FU( (4) , transformations H , and The learning problem for MFAs is to estimate the centers G 3 of sampling noise levels I of these linear processes, as well as the prior probabilities data from different parts of the manifold. Parameter estimation in MFAs can be handled by an Expectation-Maximization (EM) algorithm[5] that attempts to maximize the logprobability, , averaged over training examples. \ VXWY Z H -[ H :\/ \ \ E ba H [ H \ , Note that the parameter space of this model exhibits an invariance: taking where are orthogonal matrices ( ), does not change the marginal dis. The transformations correspond to arbitrary rotations and tribution, reflections of the local coordinates in each linear model. The objective function for the EM algorithm is unchanged by these transformations. Thus, maximum likelihood estimation in MFAs does not favor any particular alignment; instead, it produces models whose internal representations change unpredictably as one traverses connected paths on the manifold. Can we encourage models whose local coordinate systems are aligned in a consistent way? 2]_ ^`] 3 Global Coordination Suppose the data lie near a smooth manifold with a locally flat (developable) structure. Then there exist a single set of ?global coordinates? which parametrize the manifold c 1 Although in principle each neighborhood could have a different prior on its local coordinates, without loss of generality we have made the standard assumption that is the same for all settings of and absorbed the shape of each local Gaussian model into the matrices . k d/egfih*j k?l m2h hidden variables s,z Figure 1: Graphical model for globally coordinated MFAs. Al- g though global coordinates are unobserved, they affect the learning through a regularization term. After learning, inferences about the global variables are made by computing posterior distributions,   . Likewise, data can easily be generated by sampling from . All these operations are particthe conditional distribution,  ularly tractable due to the conditional independencies of the model. d/e j .l x global coordinates data d/e Zj l everywhere. Furthermore, to a good approximation, these global coordinates can be related to the local coordinates of different neighborhoods (in their region of validity) by linear 2 transformations:   (5) What does it mean to say that the coordinates provide a global parameterization of the manifold? Intuitively, if a data point belongs to overlapping neighborhoods, then the global coordinates computed from their local coordinate systems, given by eq. (5), should agree. We can formalize this ?global coordination? of different local models by treating the coordinates as unobserved variables and incorporating them into the probabilistic model: c 6    c  S ( c   ?  c    $ A c C     C )  (6) (Here we posit a deterministic relationship between local and global coordinates, although it is possible to add noise to this mapping as well.) The globally coordinated MFA is represented by the graphical model in Fig. 1. We can appeal to its conditional independencies to make other useful inferences. In particular:  c   R:   c   '  (7) ] R R c           ' P 6   '  c   R: ?( (8)    with Now, if two or more mixture components?say, K and < ?explain a data point non-negligible probability, then the posterior distributions for the global  coordinates of this induced by eq. (8), should be nearly identical: that is, c   R: K   c data   point, < . Toasenforce we need to penalize models whose  this   'criterion  given byofeq.agreement, posterior distributions c (8) are multimodal, since multiple modes only arise when different mixture components to inconsistent global coordinates. While  give   ' rise directly penalizing multimodality of c is difficult, a penalty which encourages consistency can be easily incorporated into the learning algorithm. We introduce  c a family T  ' , ofto unimodal distributions over both and , and encourage the true posteriors, c  :   ' , of this family. be close to some member, c         Developing this idea further, we introduce a new objective function for unsupervised learning in MFAs. The new objective function incorporates a regularizer to encourage the global consistency of local models:   P VXWY  ' C  P   c   c T  i XV WY  c : T   i  c    (9) The first term in this objective function computes the log-probability of the data. The second term computes a sum of Kullback-Leibler (KL) divergences; these are designed to h 2 Without loss of generality, the matrices  can be taken to be symmetric and positive-definite, by exploiting the polar factorization and absorbing reflection and rotation into the local coordinate systems. (In practice, though, it may be easier to optimize the objective function without constraining the matrices to be of this form.) In the experiments reported below, we have further restricted them to be diagonal. Together, then, the coordination matrices  and vectors account for an axis-aligned scaling and uniform translation between the global and local coordinate systems. h h penalize MFAs whose posterior distributions over global coordinates are not unimodal. The twin goals of density estimation and manifold learning in MFAs are pursued by attempting  to balance these terms in the objective function. The factor controls the tradeoff between  density modeling and global coordination: as only strict invariances (which do not affect likelihood) are exploited in order to achieve submodel agreement. In what follows  arbitrarily; further optimization is possible. we have set [  % The most convenient way to parameterize the family of unimodal distributions is a factorized form involving a Gaussian density and a multinomial:  c   ' T  '  c   '  c   '  6T  '2  (10)  T  ' in eq. (10) factorizes over  and c , implying that? Note that the distribution c according to this family of models?the coordinate is independent of the mixture  . global  c    isc Gaussian, component given the data point Also, andc :thus T  i unimodal. These are exactly the constraints we wish to impose on the posterior  , and mixture weights. At  eachare iteration of learning, the means c , covariance matrices  determined separately for each data point, so as to maximize the in :T objective   , best function c eq. (9): this amounts to computing the unimodal distributions, matched to  :   ' . the true posterior distributions, c  c T  '2           4 Learning Algorithm Latent variable models are traditionally estimated by maximum likelihood or Bayesian methods whose objective functions do not reward the interpretability of their internal representations. Note how the goal of developing more useful internal representations has changed the learning problem in a fundamental way. Now we have additional ?coordina  tion? parameters?the offsets and weights ?that must also be learned from examples. We also have auxiliary parameters for each data point?the means , covariance matri , and mixture weights ?that determine the target distributions,  . All ces these parameters, as well as the MFA model parameters , must be chosen to ?stitch together? the local coordinates systems in a smooth way and to learn internal representations easily coordinated by the local-to-global mapping in eq. (6).   c  T  ' #3  H  G  I , c Optimization of the objective function in eq. (9) is reminiscent of so-called ?variational? methods for approximate learning[7]. In these methods, an approximation to an exact (but intractable) posterior distribution is fitted by minimizing a KL divergence between the two distributions. The auxiliary parameters of the approximating distribution are known as variational parameters. Our objective function illustrates an unexpected application of such variational methods: not to perform approximate inference in intractable probabilistic models, but to learn more useful internal representations in tractable ones. We introduce the  unimodal and factorized distributions  to regularize the multimodal distributions . Penalizing the KL divergence between these distributions lifts a degeneracy in the model?s parameter space and favors local linear models that can be globally aligned.  c T  '  c T   4.1 Computing and optimizing the objective function Evaluating the objective function in eq. (9) requires a sum and integral over the latent variables of the model. These operations are simplified by rewriting the objective function as:  c :   ' M C V W Y  c T  ' S VXWY  R c  N ( (11)  :T  ' makes it straightforward to perform the The factored form of the distributions c  and required sums and integrals. The final result is a simple form in terms of entropies   P   c      energies   O  C   %    C  S & VXWY V W Y associated with the  th data point:  P    (12) ]& V W Y &7. (13) % % %   & c E  c  S & E I 9 K   C c E   E H E I 9 K  S & M   N S &% VXWY  I 8  S V W Y   C VXWY 3R S &S ] VXWY 6&87.? (14) C         G where we haveC introduced simplifying notation for the vector differences  a      9K  S H E I 9 K H 9 K .  c   and c and the local precision matrices       Iteratively maximizing the objective function by coordinate ascent now leads to a learning algorithm of the same general style as EM. # c   R  ,   %   P    c   P       % $ 9!9#" &" $ ( (15)      "  S 9 K  E E where  43 . NoticeH thatI  and. These  equations can be solved by iteration with initialization only need to be computed once before iterating the fixed point equations. The objective function is completely invariant  C to translation   E ). To and rescaling of c and (since , and c appear only in the form c 4.2 E-step Maximizing the objective function, eq. (9), with respect to the regularizing parameters  (and subject to the constraint  ) leads to the fixed point equations: 9K        remove this degeneracy, after solving the equations above we further constrain the global coordinates to have mean zero and unit variance in each direction. These constraints are enforced without changing the value of the objective function by simply translating the  offsets ' and rescaling the diagonal matrices . 4.3 M-step The M-step consists of maximizing the objective function, eq. (9), with respect to the generative parameters. Let us denote the updated parameter estimates by ( model ( )( + * ,*  ( . Letting , the M-step updates for the first three  of these are: ( .!/ * $  * ,(16) %$ # 3R    G  H  I  * ,  3 * P 9 K P  c G 9K P  Z R( The remaining updates, shown, are given in terms of  *   to beC performed * 0*   in theC order * . updated *  +*  difference vectors  , the correlations G c c 1  c E , E  *  *  N  S M  c c . and the variances 2 ( 9 K  H 1 2 (17) 3 ( I 465 - 9 K P  B 3  *  C H ( 9 K c * 4 <5 S 3 H ( 9K    E H ( E 465?F (18) ( 9 K a S H E I 9 K H  9 8K 7 E * S H E I ( 9 K 1 + 9 2 9K (19) At the optimum, the coordination weights satisfy an algebraic Riccati equation which          can be solved by iterating the update shown above. (Such equations can also be solved by much more sophisticated methods well known in the engineering community. Most approaches involve inverting the previous value of which may be expensive for full matrices but is fast in our diagonal implementation.) Figure 2: Global coordination of local linear models. (left) A model trained using maximum likelihood, with the arrows indicating the direction of increase for each factor analyzer?s local coordinate system. (right) A coordinated model; arrows indicate the direction in the data space corresponding to increasing the global coordinate as inferred by the algorithm. The ellipses show the one standard deviation contour of the density of each analyzer. 5 Experiments We have tested our model on simple synthetic manifolds whose structure is known as well as on collections of images of handwritten digits and faces. Figure 2 illustrates the basic concept of coordination, as achieved by our learning rule. In the coordinated model, the global coordinate always points in the same direction along the data manifold, as defined  by the composition of the transformations and . In the model trained with maximum likelihood, the density is well captured but each local latent variable has a random orientation along the manifold. H  We also applied the algorithm to collections of images of handwritten digits and of faces. The representation of was an unprocessed vector  of raw 8-bit grayscale pixel  intensities for each image (of dimensionality 256 for the digits and 560 for the faces.) The MFAs had 64 local models and the global coordinates were two dimensional. After training, the coordinated MFAs had learned a smooth, continuous mapping from the plane to images of digits or of faces. This allows us both to infer a two-dimensional location given any image by computing and to generate new images from any point in the plane . (Precisely what we wanted from the magic box.) In general, both by computing of these conditional distributions have the form of a mixture of Gaussians. Figure 3 shows the inferred global coordinates (i.e. the means of the unimodal distributions  ) of the training points after the last iteration of training as well as examples of new images from the generative model, created by evaluating the mean of along straight line paths in the global coordinate space. In the case of digits, it seems as though our models have captured tilt/shape and identity and represented them as the two axes of the space; in the case of the faces the axes seem to capture pose and expression. (For the faces, the final space was rotated by hand to align interpretable directions with the coordinate axes.)  c  % ^ % & ^ &  c  2 c  c    c  c c As with all EM algorithms, the coordinated MFA learning procedure is susceptible to local optima. Crucial to the success of our experiments is a good initialization, which was provided by the Locally Linear Embedding algorithm[9]. We clamped equal to the embedding coordinate provided by LLE and to a small value and trained until convergence (typically 30-100 iterations). Then we proceeded with training using the full EM equations to update , again until convergence (usually 5-10 more iterations). Note, however, that LLE and other embedding algorithms such as Isomap[10] are themselves unsupervised, so the overall procedure, including this initial phase, is still unsupervised.  c c 6 Discussion Mixture models provide a simple way to approximate the density of high dimensional data that lies on or near a low dimensional manifold. However, their hidden representations do not make explicit the relationship between dissimilar data vectors. In this paper, we have shown how to learn global coordinates that can act as an encapsulating interface, so that other parts of a learning system do not need to interact with the individual components of a mixture. This should improve generalization as well as facilitate the propagation and exchange of information when these models are incorporated into a larger (perhaps Figure 3: Automatically constructed two dimensional global parameterizations of manifolds of digits and faces. Each plot shows the global coordinate space discovered by the unsupervised algorithm; points indicate the inferred means  for each training item at the end of learning. The image stacks on the borders are not from the training set but are generated from the model itself and represent the mean of the predictive distribution  at the corresponding open circles (sampled along the straight lines in the global space). d/e Zj l The models provide both a two degree-of-freedom generator for complex images  via as well as a pose/slant recognition system via   . d/e Zj l d/e j .l For the handwritten digits, the training set consisted of 1100 examples of the digit ?2? (shown as crosses above) mixed with 1100 examples of ?3?s (shown as triangles). The digits are from the NIST dataset, digitized at 16x16 pixels. For the faces, we used 2000 images of a single person with various poses and expressions taken from consecutive frames of a video digitized at 20x20 pixels. Brendan Frey kindly provided the face data. hierarchical) architecture for probabilistic reasoning. Two variants of our purely unsupervised proposal are possible. The first is to use an embedding algorithm (such as LLE or Isomap) not only as an initialization step but to provide clamped values for the global coordinates. While this supervised approach may work in practice, unsupervised coordination makes clear the objective function that is being opti- Figure 4: A situation in which an un-coordinated mixture model?trained to do density estimation?cannot be ?postcoordinated?. Noise has caused one of the local density models to orient orthogonal to the manifold. In globally coordinated learning, there is an additional pressure to align with neighbouring models which would force the local model to lie in the correct subspace. mized, which unifies the goals of manifold learning and density estimation. Another variant is to train an unsupervised mixture model (such as a MFA) using a traditional maximum likelihood objective function and then to ?post-coordinate? its parameters by applying local reflections/rotations and translations to create global coordinates. As illustrated in figure 4, however, this two-step procedure can go awry because of noise in the original training set. When both density estimation and coordination are optimized simultaneously there is extra pressure for local experts to fit the global structure of the manifold. Our work can be viewed as a synthesis of two long lines of research in unsupervised learning. In the first are efforts at learning the global structure of nonlinear manifolds [1, 4, 9, 10]; in the second are efforts at developing probabilistic graphical models for reasoning under uncertainty[5, 6, 7]. Our work proposes to model the global coordinates on manifolds as latent variables, thus attempting to combine the representational advantages of both frameworks. It differs from embedding by providing a fully probabilistic model valid away from the training set, and from work in generative topographic mapping[2] by not requiring a uniform discretized gridding of the latent space. Moreover, by extending the usefulness of mixture models,it further develops an architecture that has already proved quite powerful and enormously popular in applications of statistical learning. Acknowledgements We thank Mike Revow for sharing his unpublished work (at the University of Toronto) on coordinating mixtures, and Zoubin Ghahramani, Peter Dayan, Jakob Verbeek and two anonymous reviewers for helpful comments and corrections. References [1] D. Beymer & T. Poggio. Image representations for visual learning. pringerScience 272 (1996). [2] C. Bishop, M. Svensen, and C. Williams. GTM: The generative topographic mapping. Neural Computation 10 (1998). [3] C. Bregler & S. Omohundro. Nonlinear image interpolation using manifold learning. Advances in Neural Information Processing Systems 7 (1995). [4] D. DeMers & G.W. Cottrell. Nonlinear dimensionality reduction. Advances in Neural Information Processing Systems 5 (1993). [5] Ghahramani, Z. and Hinton, G. The EM algorithm for mixtures of factor analyzers. University of Toronto Technical Report CRG-TR-96-1 (1996). [6] Hinton, G., Dayan, P., and Revow, M. Modeling the manifolds of images of handwritten digits. IEEE Transactions on Neural Networks 8 (1997). [7] M. Jordan, Z. Ghahramani, T. Jaakkola, and L. Saul. An introduction to variational methods for graphical models. Machine Learning 37(2) (1999). [8] N. Kambhatla and T. K. Leen. Dimension reduction by local principal component analysis. Neural Computation 9 (1997). [9] S. T. Roweis & L. K. Saul. Nonlinear dimensionality reduction by locally linear embedding. Science 290 (2000). [10] J. B. Tenenbaum, V. de Silva, and J. C. Langford. A global geometric framework for nonlinear dimensionality reduction. Science 290 (2000).
2082 |@word proceeded:1 loading:1 seems:1 open:1 covariance:3 simplifying:1 pressure:2 tr:1 reduction:4 initial:1 contains:1 document:1 must:2 reminiscent:1 cottrell:1 shape:2 wanted:1 remove:1 treating:1 designed:1 update:4 interpretable:1 plot:1 implying:1 pursued:1 generative:4 item:1 parameterization:2 plane:2 awry:1 parameterizations:1 toronto:3 traverse:2 location:4 simpler:1 along:4 vxw:7 constructed:1 consists:1 combine:1 multimodality:1 introduce:3 themselves:1 nor:1 discretized:1 globally:4 automatically:2 increasing:1 provided:3 underlying:1 matched:1 notation:1 factorized:2 moreover:1 what:3 unobserved:2 transformation:5 act:1 usefully:1 exactly:1 control:1 normally:1 unit:1 appear:1 arguably:1 positive:1 negligible:1 before:1 frey:1 local:33 tends:1 engineering:1 xv:1 despite:1 opti:1 path:4 interpolation:1 black:1 initialization:3 factorization:1 averaged:1 practice:2 tribution:1 definite:1 differs:1 digit:10 procedure:3 convenient:1 integrating:1 zoubin:1 cannot:1 close:1 applying:1 optimize:1 deterministic:1 reviewer:1 center:2 maximizing:3 straightforward:1 go:1 williams:1 factored:1 rule:1 array:1 submodel:1 regularize:1 his:1 embedding:6 coordinate:45 traditionally:1 updated:2 imagine:1 suppose:1 target:1 exact:1 neighbouring:1 agreement:2 expensive:1 recognition:1 bec:1 observed:2 mike:1 solved:3 capture:2 parameterize:2 region:1 connected:2 intuition:1 reward:1 trained:4 solving:1 predictive:1 purely:1 completely:1 triangle:1 easily:3 joint:1 multimodal:2 represented:5 various:1 gtm:1 regularizer:3 train:1 fast:1 describe:2 lift:1 neighborhood:7 whose:7 quite:1 widely:1 larger:1 say:2 favor:2 topographic:2 itself:1 final:2 advantage:1 aligned:4 deduces:1 riccati:1 achieve:1 roweis:2 representational:1 description:1 exploiting:1 convergence:2 optimum:2 extending:1 produce:1 rotated:1 svensen:1 pose:4 eq:9 solves:1 strong:1 auxiliary:2 indicate:2 direction:5 posit:1 correct:1 translating:1 exchange:1 generalization:1 anonymous:1 bregler:1 crg:1 correction:1 lawrence:1 mapping:5 kambhatla:1 consecutive:1 estimation:6 polar:1 coordination:11 create:1 sensor:1 gaussian:4 always:1 varying:1 factorizes:1 jaakkola:1 ax:3 likelihood:7 brendan:1 helpful:2 inference:4 dayan:2 typically:1 hidden:5 favoring:1 pixel:7 overall:1 orientation:1 proposes:1 fairly:1 marginal:2 equal:1 construct:1 once:1 nicely:1 sampling:2 identical:1 represents:1 unsupervised:10 nearly:1 report:1 ofto:1 develops:1 simultaneously:1 divergence:4 individual:1 phase:1 attempt:1 freedom:1 highly:2 alignment:1 mixture:17 isc:1 egfih:1 fu:1 encourage:3 integral:2 poggio:1 orthogonal:2 circle:1 fitted:1 modeling:2 maximization:1 deviation:1 neutral:1 uniform:2 usefulness:1 characterize:1 reported:1 synthetic:1 person:1 density:10 fundamental:1 probabilistic:7 together:2 continuously:1 synthesis:1 again:1 lever:3 expert:1 style:1 rescaling:2 account:1 de:1 twin:1 coordinated:9 satisfy:1 caused:1 tion:1 break:1 performed:1 analyze:1 complicated:1 variance:2 likewise:1 ensemble:2 correspond:1 raw:2 bayesian:1 handwritten:4 unifies:1 straight:2 explain:1 sharing:1 against:1 energy:1 associated:1 degeneracy:3 sampled:2 demers:1 dataset:1 proved:1 popular:1 dimensionality:5 formalize:1 sophisticated:1 supervised:1 leen:1 box:5 though:3 generality:2 furthermore:1 correlation:2 until:2 hand:1 langford:1 nonlinear:6 overlapping:1 propagation:1 thati:1 mode:1 perhaps:1 scientific:1 facilitate:1 validity:1 requiring:1 true:2 isomap:2 concept:1 consisted:1 regularization:1 symmetric:1 leibler:2 iteratively:1 illustrated:1 encourages:1 criterion:1 omohundro:1 reflection:3 interface:1 silva:1 reasoning:3 image:23 variational:5 regularizing:2 rotation:3 absorbing:1 multinomial:1 qp:1 tilt:1 composition:1 slant:1 consistency:2 analyzer:5 had:2 add:1 align:2 posterior:8 optimizing:1 belongs:1 arbitrarily:1 success:1 exploited:1 captured:2 fortunately:1 additional:2 impose:1 spectrogram:1 determine:1 maximize:2 multiple:1 unimodal:7 full:2 infer:1 smooth:3 technical:1 cross:2 long:1 post:1 manipulate:1 ellipsis:1 verbeek:1 involving:1 basic:1 variant:2 expectation:1 iteration:5 represent:3 achieved:2 penalize:2 proposal:1 background:1 separately:1 crucial:1 extra:1 strict:1 ascent:1 induced:1 subject:1 comment:1 member:1 inconsistent:1 incorporates:1 seem:1 jordan:1 near:5 constraining:1 affect:2 fit:1 pennsylvania:1 architecture:2 idea:1 tradeoff:1 unprocessed:1 expression:4 handled:1 effort:2 penalty:1 peter:1 algebraic:1 speech:1 logprobability:1 useful:5 iterating:2 clear:1 involve:1 amount:1 locally:6 tenenbaum:1 generate:1 exist:1 zj:3 estimated:1 coordinating:1 discrete:2 independency:2 changing:1 penalizing:2 ce:1 rewriting:1 sum:3 enforced:1 orient:1 parameterized:2 everywhere:1 uncertainty:1 powerful:1 family:4 scaling:1 bit:1 matri:1 constraint:3 precisely:1 constrain:1 flat:1 generates:1 attempting:2 mized:1 bdc:1 department:2 developing:3 according:1 describes:1 em:6 sam:1 intuitively:1 restricted:1 invariant:1 taken:2 equation:7 agree:1 r3:1 encapsulating:1 letting:1 tractable:3 end:1 parametrize:1 operation:3 gaussians:1 hierarchical:1 away:1 appropriate:1 gridding:1 original:1 assumes:2 remaining:1 graphical:4 dial:3 exploit:1 ghahramani:3 approximating:1 unchanged:1 objective:21 already:1 diagonal:3 traditional:1 exhibit:1 subspace:1 thank:1 manifold:32 index:1 relationship:3 providing:1 balance:1 minimizing:1 difficult:2 susceptible:1 x20:1 rise:1 ba:1 magic:2 implementation:1 perform:3 nist:1 curved:1 situation:1 hinton:3 incorporated:2 digitized:2 frame:1 discovered:1 stack:1 jakob:1 arbitrary:1 community:1 intensity:4 inferred:3 introduced:1 inverting:1 unpublished:1 required:1 kl:3 optimized:1 learned:2 usually:2 below:1 interpretability:1 including:1 video:1 mfa:15 force:1 improve:1 axis:1 created:1 prior:3 geometric:1 acknowledgement:1 loss:2 fully:1 mixed:1 geoffrey:1 generator:1 degree:1 consistent:2 principle:1 translation:3 changed:1 last:1 dis:1 lle:3 saul:3 face:21 taking:1 distributed:1 dimension:1 evaluating:2 valid:1 contour:1 computes:2 collection:6 made:2 simplified:1 transaction:1 approximate:4 kullback:2 gene:1 global:40 robotic:1 summing:1 grayscale:1 continuous:3 latent:6 un:1 learn:5 interact:1 complex:1 domain:1 kindly:1 arrow:2 border:1 noise:5 arise:1 fig:1 enormously:1 x16:1 precision:1 wish:1 explicit:1 lie:6 clamped:2 perceptual:2 bishop:1 appeal:1 offset:2 intractable:3 intrinsic:2 exists:1 incorporating:1 adding:1 illustrates:3 easier:1 smoothly:2 entropy:1 simply:1 beymer:1 absorbed:1 visual:1 unexpected:2 stitch:1 applies:1 conditional:4 goal:4 identity:1 viewed:1 revow:2 change:3 determined:1 unpredictably:1 principal:1 called:1 invariance:2 indicating:1 select:1 internal:8 support:1 dissimilar:1 tested:1
1,188
2,083
The Method of Quantum Clustering David Horn and Assaf Gottlieb School of Physics and Astronomy Raymond and Beverly Sackler Faculty of Exact Sciences Tel Aviv University, Tel Aviv 69978, Israel Abstract We propose a novel clustering method that is an extension of ideas inherent to scale-space clustering and support-vector clustering. Like the latter, it associates every data point with a vector in Hilbert space, and like the former it puts emphasis on their total sum, that is equal to the scalespace probability function. The novelty of our approach is the study of an operator in Hilbert space, represented by the Schr?odinger equation of which the probability function is a solution. This Schr?odinger equation contains a potential function that can be derived analytically from the probability function. We associate minima of the potential with cluster centers. The method has one variable parameter, the scale of its Gaussian kernel. We demonstrate its applicability on known data sets. By limiting the evaluation of the Schr?odinger potential to the locations of data points, we can apply this method to problems in high dimensions. 1 Introduction Methods of data clustering are usually based on geometric or probabilistic considerations [1, 2, 3]. The problem of unsupervised learning of clusters based on locations of points in data-space, is in general ill defined. Hence intuition based on other fields of study may be useful in formulating new heuristic procedures. The example of [4] shows how intuition derived from statistical mechanics leads to successful results. Here we propose a model based on tools that are borrowed from quantum mechanics. We start out with the scale-space algorithm of [5] that uses a Parzen-window estimator of the probability distribution based on the data. Using a Gaussian kernel, one generates from the data points in a Euclidean space of dimension  a probability distribution given by, up to an overall normalization, the expression      (1) where are the data points. It seems quite natural [5] to associate maxima of this function with cluster centers. The same kind of Gaussian kernel was the basis of another method, Support Vector Clustering (SVC) [6], associating the data-points with vectors in an abstract Hilbert space. Here we will also consider a Hilbert space, but, in contradistinction with kernel methods where the Hilbert space is implicit, here we work with a Schr?odinger equation that serves as the basic framework of the Hilbert space. Our method was introduced in [7] and is further expanded in this presentation. Its main emphasis is on the Schr?odinger potential, whose minima will determine the cluster centers. This potential is part of the Schr?odinger equation that is a solution of. 2 The Schr?odinger Potential We define[7] the Schr?odinger equation                (2)    is a solution, or eigenstate.1 The simplest case is that of a single Gaussian, for which   Then it turns out that     . This when represents a single point at .     quadratic function, whose center lies at  , is known as the harmonic potential in quantum  is the lowest possible eigenvalue of , mechanics (see, e.g., [8]). Its eigenvalue hence the Gaussian function is said to describe the ground state of .    and one searches for solutions, Conventionally, in  quantum mechanics, one is given           or eigenfunctions, we have already , asdetermined by the data points, we . Here,  whose   . This can ask therefore for the  solution is the given be easily obtained through                  is still left undefined. For this purpose we require This sets the value of and determines        "!$#                    (3)  to be positive definite, i.e. min =0. (4) uniquely. Using Eq. 3 it is easy to prove that %"& ('   (5) 3 2D Examples 3.1 Crab Data To show the power of our new method we discuss the crab data set taken from Ripley?s book [9]. This data set is defined over a five-dimensional parameter space. When analyzed in terms of the 2nd and 3rd principal components of the correlation matrix one observes a nice separation of the 200 instances into their four classes. We start therefore with this problem as our first test case. In Fig. 1 we show the data as well as the Parzen probability distribution    using the width parameter . It is quite obvious that this width is not small enough to deduce the correct clustering according to the approach of [5]. Nonetheless, the potential displayed in Fig. 2 shows the required four minima for the same width parameter. Thus we conclude that the necessary information is already available. One needs, however, the quantum clustering approach, to bring it out.  1 + + ,  *) - . (the Hamiltonian) and (potential energy) are conventional quantum mechanical operators, rescaled so that depends on one parameter, . is a (rescaled) energy eigenvalue in quantum mechanics. 2 1 2 1 0 0 ?1 ?2 ?1 ?3 ?2 PC3 PC2 Figure 1: A plot of Roberts? probability distribution for Ripley?s crab data [9] as defined over the 2nd and 3rd principal components of the correlation matrix. Using a Gaussian width of we observe only one maximum. Different symbols label the four classes of data.  ) 1.2 1 V/E 0.8 0.6 0.4 0.2 2 0 2 1 1 0 0 ?1 PC2 ?2 ?1 ?3 ?2 PC3  Figure 2: A plot of the Schr?odinger potential for the same problem as Fig. 1. Here we clearly see the required four minima. The potential is plotted in units of .  Note in Fig. 2 that the potential grows quadratically outside the domain over which the data are located. This is a general property of Eq. 3. sets the relevant scale over which one may look for structure potential. If the width is decreased more structure is to of the, two be expected. Thus, for more minima appear, as seen in Fig. 3. Nonetheless, they lie high and contain only a few data points. The major minima are the same as in Fig. 2.   3.2 Iris Data Our second example consists of the iris data set [10], which is a standard benchmark obtainable from the UCI repository [11]. Here we use the first two principal components to define the  two dimensions in which we apply our method. Fig. 4, which shows the case for , provides an almost perfect separation of the 150 instances into the three classes into which they should belong.  %  4 Application of Quantum Clustering The examples displayed in the previous section show that, if the spatial representation of the data allows for meaningful clustering using geometric information, quantum clustering (QC) will do the job. There remain, however, several technical questions to be answered: What is the preferred choice of ? How can QC be applied in high dimensions? How does one choose the appropriate space, or metric, in which to perform the analysis? We will confront these issues in this section.  4.1 Varying        In the crabs-data we find that as is decreased to , the previous minima of get deeper and two new minima are formed. However the latter are insignificant, in the sense that they lie at high values (of order ), as shown in Fig. 3. Thus, if we data-points  classify  , roughly same to clusters according to their topographic location on the surface of as for  . By the way, the wave the clustering assignment is expected for function . As is being further decreased, more acquires only one additional maximum at  and more maxima are expected in and an ever increasing number of minima (limited by ) in .            The one parameter of our problem, , signifies the distance that we probe. Accordingly we expect to find clusters relevant to proximity information of the same order of magnitude. One may therefore vary continuously and look for stability of cluster solutions, or limit oneself to relatively high values of and decide to stop the search once a few clusters are being uncovered.   4.2 Higher Dimensions In the iris problem we obtained excellent clustering results using the first two principal components, whereas in the crabs problem, clustering that depicts correctly the classification necessitates components 2 and 3. However, once this is realized, it does not harm to add the 1st component. This requires in a 3-dimensional space, spanned by the   onworking three leading PCs. Calculating a fine computational grid becomes a heavy task in high dimensions. To cut down complexity, we propose using the analytic expression of Eq. 3 and evaluating the potential on data points only. This should be good enough to give a close estimate of where the minima lie, and it reduces the complexity to irrespective of dimension. In the gradient-descent algorithm described below, we will require further computations, also restricted to well defined locations in space.   1.2 1 V/E 0.8 0.6 0.4 0.2 2 1 0 2 0 1 0 ?1 ?1 ?2 ?2 ?3 PC3   PC2 Figure 3: The potential for the crab data with displays two additional, but insignificant, minima. The four deep minima are roughly at the same locations as in Fig. 2. 1.5 1 0.5 PC2 0 ?0.5 ?1 ?1.5 ?2 ?2.5 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 PC1 %   in a space spanned by the first Figure 4: Quantum clustering of the iris data for two principal components. symbols represent the three classes. Equipotential   Different   lines are drawn at                  When locations of data points, we evaluate on a discrete   restricted . We tocanthethen  set of  points express in terms of the distance matrix as                                with chosen appropriately so that min  =0. (6) All problems that we have used as examples were such that data were given in some space, and we have exercised our freedom to define a metric, using the PCA approach, as the basis for distance calculations. The previous analysis tells us that QC can also be applied to data for which only the distance information is known. 4.3 Principal Component Metrics The QC algorithm starts from distance information. The question how the distances are calculated is another - very important - piece of the clustering procedure. The PCA approach defines a metric that is intrinsic to the data, determined by their second order statistics. But even then, several possibilities exist, leading to non-equivalent results. Principal component decomposition can be applied both to the correlation matrix and to the covariance matrix. Moreover, whitening normalization may be applied. The PCA approach that we have used is based on a whitened correlation matrix. This turns out to lead to the good separation of crab-data in PC2-PC3 and of iris-data in PC1-PC2. Since our aim was to convince the reader that once a good metric is found, QC conveys the correct information, we have used the best preprocessing before testing QC. 5 The Gradient Descent Algorithm After discovering the cluster centers we are faced with the problem of allocating the data points to the different clusters. propose using a gradient descent algorithm for this   weWe purpose. Defining define the process %              (7) letting the points reach an asymptotic fixed value coinciding with a cluster center. More sophisticated minimum search algorithms, as given in chapter 10 of [12], may be used for faster convergence. To demonstrate the results of this algorithm, as well as the application of QC to higher dimensions, we analyze the iris data in 4 dimensions. We use the original data space with only one modification: all axes are normalized to lie within a unified range of variation. The results are displayed in Fig. 5. Shown here are different windows for the four different axes, within which we display the values of the points after descending the potential surface and reaching its minima, whose values are shown in the fifth window. These results are very satisfactory, having only 5 misclassifications. Applying QC to data space without normalization of the different axes, leads to misclassifications of the order of 15 instances, similar to the clustering quality of [4].  6 Discussion In the literature of image analysis one often looks for the curve on which the Laplacian of the Gaussian filter of an image vanishes[13]. This is known as zero-crossing and serves as dim 1 1.5 1 dim 2 0.5 0 50 100 150 0 50 100 150 0 50 100 150 0 50 100 150 1 dim 3 2 1 0 dim 4 2 1 0 V/E 0.2 0.1 0 20 40 60 80 serial number 100 120 140 Figure 5: The fixed points of the four-dimensional iris problem following the gradientdescent algorithm. The show almost perfect clustering into the three families of 50 results instances each for .  %   % a measure of the image. Its analogue in the scale-space approach is where   of .segmentation Clearly each such contour can also be viewed as surrounding maxima of the probability function, and therefore representing some kind of cluster boundary, although different from the conventional one [5]. It is known that the number of such boundaries [13] is a non-decreasing function of . Note that such can be read off Fig. 4.   contours contours on the periphery of Comparison with Eq. 3 tells us that they are the this figure. Clearly they surround the data but do not give a satisfactory indication of where  the clusters are. Cluster cores are better defined by curves in this figure. One may therefore speculate that equipotential levels of may serve as alternatives to curves in future applications to image analysis.       %   % Image analysis is a 2-dimensional problem, in which differential operations have to be formulated and followed on a fine grid. Clustering is a problem that may occur in any number of dimensions. It is therefore important to develop a tool that can deal with it accordingly. Since the Schr?odinger potential, the function that plays the major role in our analysis, has minima that lie in the neighborhood of data points, we find that it suffices to evaluate it at these points. This enables us to deal with clustering in high dimensional spaces. The results, such as the iris problem of Fig. 5, are very promising. They show that the basic idea, as well as the gradient-descent algorithm of data allocation to clusters, work well. Quantum clustering does not presume any particular shape or any specific number of clusters. It can be used in conjunction with other clustering methods. Thus one may start with SVC to define outliers which will be excluded from the construction of the QC potential. This would be one example where not all points are given the same weight in the construction of the Parzen probability distribution. It may seem strange to see the Schr?odinger equation in the context of machine learning. Its usefulness here is due to the fact that the two different terms of Eq. 2 have opposite effects on the wave-function. The potential represents the attractive force that tries to concentrate the distribution around its minima. The Laplacian has the opposite effect of spreading the wave-function. In a clustering analysis we implicitly assume that two such effects exist. QC models them with the Schr?odinger equation. Its success proves that this equation can serve as the basic tool of a clustering method. References [1] A.K. Jain and R.C. Dubes. Algorithms for clustering data. Prentice Hall, Englewood Cliffs, NJ, 1988. [2] K. Fukunaga. Introduction to Statistical Pattern Recognition. Academic Press, San Diego, CA, 1990. [3] R.O. Duda, P.E. Hart and D.G. Stork. Pattern Classification. Wiley-Interscience, 2nd ed., 2001. [4] M. Blat, S. Wiseman and E. Domany. Super-paramagnetic clustering of data. Phys. Rev. Letters 76:3251-3255, 1996. [5] S.J. Roberts. Non-parametric unsupervised cluster analysis. Pattern Recognition, 30(2):261?272, 1997. [6] A. Ben-Hur, D. Horn, H.T. Siegelmann, and V. Vapnik. A Support Vector Method for Clustering. in Advances in Neural Information Processing Systems 13: Proceedings of the 2000 Conference Todd K. Leen, Thomas G. Dietterich and Volker Tresp eds., MIT Press 2001, pp. 367?373. [7] David Horn and Assaf Gottlieb. Algorithm for Data Clustering in Pattern Recognition Problems Based on Quantum Mechanics. Phys. Rev. Lett. 88 (2002) 018702. [8] S. Gasiorowicz. Quantum Physics. Wiley 1996. [9] B. D. Ripley Pattern Recognition and Neural Networks. Cambridge University Press, Cambridge UK, 1996. [10] R.A. Fisher. The use of multiple measurements in taxonomic problems. Annual Eugenics, 7:179?188, 1936. [11] C.L. Blake and C.J. Merz. UCI repository of machine learning databases, 1998. [12] W. H. Press, S. A. Teuklosky, W. T. Vetterling and B. P. Flannery. Numerical Recipes - The Art of Scientific Computing 2nd ed. Cambridge Univ. Press, 1992. [13] A. L. Yuille and T. A. Poggio. Scaling theorems for zero crossings. IEEE Trans. Pattern Analysis and Machine Intelligence PAMI-8, 15-25, 1986.
2083 |@word repository:2 faculty:1 seems:1 duda:1 nd:4 decomposition:1 covariance:1 contains:1 uncovered:1 paramagnetic:1 numerical:1 shape:1 analytic:1 enables:1 plot:2 intelligence:1 discovering:1 accordingly:2 hamiltonian:1 core:1 provides:1 location:6 five:1 differential:1 prove:1 consists:1 assaf:2 interscience:1 expected:3 roughly:2 mechanic:6 decreasing:1 window:3 increasing:1 becomes:1 moreover:1 lowest:1 israel:1 what:1 kind:2 unified:1 astronomy:1 nj:1 every:1 uk:1 unit:1 appear:1 positive:1 before:1 todd:1 limit:1 cliff:1 pami:1 emphasis:2 limited:1 range:1 horn:3 testing:1 definite:1 procedure:2 get:1 close:1 operator:2 put:1 context:1 applying:1 prentice:1 descending:1 conventional:2 equivalent:1 center:6 qc:10 estimator:1 spanned:2 stability:1 variation:1 limiting:1 construction:2 play:1 diego:1 exact:1 us:1 associate:3 crossing:2 recognition:4 located:1 cut:1 database:1 role:1 rescaled:2 observes:1 intuition:2 vanishes:1 complexity:2 serve:2 yuille:1 basis:2 necessitates:1 easily:1 represented:1 chapter:1 surrounding:1 univ:1 jain:1 describe:1 tell:2 outside:1 neighborhood:1 quite:2 heuristic:1 whose:4 statistic:1 topographic:1 eigenvalue:3 indication:1 propose:4 relevant:2 uci:2 recipe:1 convergence:1 cluster:17 perfect:2 ben:1 develop:1 dubes:1 school:1 borrowed:1 eq:5 job:1 concentrate:1 correct:2 filter:1 require:2 suffices:1 extension:1 proximity:1 crab:7 around:1 ground:1 hall:1 blake:1 major:2 vary:1 purpose:2 label:1 spreading:1 exercised:1 tool:3 mit:1 clearly:3 gaussian:7 aim:1 super:1 reaching:1 varying:1 volker:1 conjunction:1 derived:2 ax:3 sense:1 dim:4 vetterling:1 overall:1 issue:1 ill:1 classification:2 spatial:1 art:1 equal:1 field:1 once:3 having:1 represents:2 look:3 unsupervised:2 future:1 inherent:1 few:2 freedom:1 englewood:1 possibility:1 evaluation:1 analyzed:1 undefined:1 pc:1 allocating:1 necessary:1 poggio:1 euclidean:1 plotted:1 instance:4 classify:1 wiseman:1 assignment:1 signifies:1 applicability:1 usefulness:1 successful:1 convince:1 st:1 probabilistic:1 physic:2 off:1 parzen:3 continuously:1 choose:1 book:1 leading:2 potential:19 speculate:1 depends:1 piece:1 try:1 analyze:1 start:4 wave:3 formed:1 beverly:1 presume:1 reach:1 phys:2 ed:3 nonetheless:2 energy:2 pp:1 obvious:1 conveys:1 stop:1 ask:1 hur:1 hilbert:6 segmentation:1 obtainable:1 sophisticated:1 higher:2 coinciding:1 leen:1 implicit:1 correlation:4 defines:1 quality:1 scientific:1 aviv:2 grows:1 effect:3 dietterich:1 contain:1 normalized:1 former:1 analytically:1 hence:2 read:1 excluded:1 satisfactory:2 deal:2 attractive:1 width:5 uniquely:1 acquires:1 iris:8 equipotential:2 demonstrate:2 scalespace:1 bring:1 image:5 harmonic:1 consideration:1 novel:1 svc:2 stork:1 belong:1 measurement:1 surround:1 cambridge:3 rd:2 grid:2 surface:2 whitening:1 deduce:1 add:1 periphery:1 success:1 seen:1 minimum:16 additional:2 novelty:1 determine:1 multiple:1 reduces:1 technical:1 faster:1 academic:1 calculation:1 hart:1 serial:1 laplacian:2 basic:3 whitened:1 confront:1 metric:5 kernel:4 normalization:3 represent:1 whereas:1 fine:2 decreased:3 appropriately:1 pc3:4 eigenfunctions:1 seem:1 easy:1 enough:2 misclassifications:2 associating:1 opposite:2 idea:2 domany:1 oneself:1 expression:2 pca:3 deep:1 useful:1 simplest:1 exist:2 correctly:1 discrete:1 express:1 four:7 drawn:1 sum:1 letter:1 taxonomic:1 almost:2 reader:1 decide:1 family:1 strange:1 separation:3 pc2:6 scaling:1 sackler:1 followed:1 display:2 quadratic:1 annual:1 occur:1 generates:1 answered:1 min:2 formulating:1 fukunaga:1 expanded:1 relatively:1 according:2 remain:1 rev:2 modification:1 outlier:1 restricted:2 taken:1 equation:8 turn:2 discus:1 letting:1 serf:2 available:1 operation:1 apply:2 observe:1 probe:1 appropriate:1 alternative:1 original:1 thomas:1 clustering:28 calculating:1 siegelmann:1 prof:1 already:2 question:2 realized:1 parametric:1 said:1 gradient:4 distance:6 robert:2 perform:1 benchmark:1 descent:4 displayed:3 defining:1 ever:1 schr:12 pc1:2 david:2 introduced:1 required:2 mechanical:1 quadratically:1 trans:1 eugenics:1 usually:1 below:1 pattern:6 analogue:1 power:1 natural:1 force:1 representing:1 irrespective:1 conventionally:1 tresp:1 raymond:1 faced:1 nice:1 geometric:2 literature:1 asymptotic:1 expect:1 allocation:1 heavy:1 deeper:1 fifth:1 curve:3 dimension:10 calculated:1 evaluating:1 boundary:2 contour:3 quantum:13 lett:1 preprocessing:1 san:1 preferred:1 implicitly:1 harm:1 conclude:1 ripley:3 search:3 promising:1 ca:1 tel:2 excellent:1 domain:1 main:1 fig:12 depicts:1 wiley:2 lie:6 down:1 gradientdescent:1 theorem:1 specific:1 symbol:2 insignificant:2 intrinsic:1 vapnik:1 magnitude:1 flannery:1 contradistinction:1 determines:1 viewed:1 presentation:1 formulated:1 fisher:1 determined:2 gottlieb:2 principal:7 total:1 odinger:12 merz:1 meaningful:1 support:3 latter:2 evaluate:2
1,189
2,084
Intransitive Likelihood-Ratio Classifiers Jeff Bilmes and Gang Ji Department of Electrical Engineering University of Washington Seattle, WA 98195-2500 bilmes,gji  @ee.washington.edu Marina Meil?a Department of Statistics University of Washington Seattle, WA 98195-4322 mmp@stat.washington.edu Abstract In this work, we introduce an information-theoretic based correction term to the likelihood ratio classification method for multiple classes. Under certain conditions, the term is sufficient for optimally correcting the difference between the true and estimated likelihood ratio, and we analyze this in the Gaussian case. We find that the new correction term significantly improves the classification results when tested on medium vocabulary speech recognition tasks. Moreover, the addition of this term makes the class comparisons analogous to an intransitive game and we therefore use several tournament-like strategies to deal with this issue. We find that further small improvements are obtained by using an appropriate tournament. Lastly, we find that intransitivity appears to be a good measure of classification confidence. 1 Introduction An important aspect of decision theory is multi-way pattern classification whereby one must determine the class  for a given data vector  that minimizes the overall risk:   argmin          where      is the loss in choosing   when the true class is  . This decision rule is provably optimal for the given loss function [3]. For the 0/1-loss functions, it is optimal to simply use the posterior probability to determine the optimal class    argmax     This procedure may equivalently be specified using a tournament style game-playing strat(egy. In this case, there is an implicit class ordering  "!#!$!#%& , and a class-pair (' and ) scoring function for an unknown sample  : )+*-, *-, (    *.,  0/21 *., *., ( 43#576   8 '9:;  8  is the log-likelihood)8ratio 43$56   '<=:>   is such that <?< @ and 1 the ) log odds. The strategy proceeds by evaluating which if positive is followed  ?  prior A   A  ) @ by and otherwise by . This continues until a ?winner? is found. Of course, the order of the classes does not matter, as the same winner is found for all permutations. In any event, this style of classification can be seen as a transitive game [5] between players who correspond to the individual classes. In this work we extend the likelihood-ratio based classification with a term, based on the Kullback-Leibler divergence [2], that expresses the inherent posterior confusability between the underlying likelihoods being compared for a given pair of players. We find that by including this term, the results of a classification system significantly improve, without changing or increasing the quantity of the estimated free model parameters. We also show how, under certain assumptions, the term can be seen as an optimal correction between the estimated model likelihood ratio and the true likelihood ratio, and gain further intuition by examining the case when the likelihoods   8 '9 are Gaussians. Furthermore, we observe that the new strategy leads to an intransitive game [5], and we investigate several strategies for playing such games. This results in further (but small) improvements. Finally, we consider the instance of intransitivity as a confidence measure, and investigate an iterative approach to further improve the correction term. Section 2 first motivates and defines our approach, and shows the conditions under which it is optimal. Section 2.1 then reports experimental results which show significant improvements where the likelihoods are hidden Markov models trained on speech data. Section 3 then recasts the procedure as intransitive games, and evaluates a variety of game playing strategies yielding further (small) error reductions. Section 3.1 attempts to better understand our results via empirical analysis, and evaluates additional classification strategies. Section 4 explores an iterative strategy for improving our technique, and finally Section 5 concludes and discusses future work. 2 Extended Likelihood-Ratio-based Classification The Kullback-Leibler (KL) divergence[2], an asymmetric measure of the distance between two probability densities, is defined as follows:        83$56   where  and  are probability densities over the same sample space. The KL-divergence is also called the average (under  ) information for discrimination in favor of  over  . For our purposes, we are interested in KL-divergence between class-conditional likelihoods   8 '9 where ' is the class number: (  '    3$56 (   8 '< (  8 '<7   8   ( One intuitive way of viewing  '  is as follows: if  ' (  is small, then (samples of to be( erroneously classified as class than when  '  is large. class ' are more likely ( ( Comparing  '   and  >'< should tell us which of ' and is more likely ( to have ( its samples mis-classified by the other model. Therefore, the difference  '       '9 , ( when positive, indicates that samples of class are more likely( to be mis-classified as class ' than samples of class ' are to be mis-classified as class (and vice-versa when ( ( the difference is negative). In other words, ' ?steals? from ( more than steals from ' when the difference is positive, thereby suggesting that class should receive aid in this case. This difference can be viewed as a form of posterior (i.e., based on the data) ?bias? *., indicating which class should receive favor over the other.1 We can adjust (the log(likelihood ratio) with this posterior bias, to obtain a new function comparing classes ' and as follows: ) *-, *-, *-, *.,  1 / 1  ! Note that this is not the normal notion of statistical bias as in model parameters.  where  is an estimate of   where *-, (    '   (  (  >'<  -* , The is positive, and in favor of ' when *., likelihood ratio is adjusted ) *-, in favor of when is negative. We then use , and when it is positive, choose class ' . The above *-, intuition does *-, not explain why such a correction factor should be used, since along with 1 is already optimal. In practice, however, we do not have access using to the true likelihood ratios but instead to an approximation that has been estimated from *., (  $ 3  5 6 training data. Let the variable be the true log-likelihood ratio,     8   < ' =  > :        *.,   (  and    3#576   8 '9:   8  be the model-based log ratio. Furthermore, let  (  '       8 '< (  8 '<7   8    3$56 be the modified KL-divergence between the class  *.,    conditional  ( models, measured *.modulo , (   the true distribution     '< , and let (resp. '      >'< . Finally, let 1 1  *., ) be the true (resp. estimated) log prior odds. Our (usable) scoring function becomes: )+*.,     *-,   0/ 1  *-,   *-, ! (1) which has an intuitive explanation similar to the above. There are certain conditions under which the above approach is theoretically justifiable. Let ( ( . us assume for now a two-class problem where ' and are the two classes, so   '9;/     A sufficient condition is   for the estimated quantities above to yield optimal performance ( for / 1  / 1 for all  .2 Since this is not the case in practice, an ' -dependent constant term  may  be added correcting for any differences as best as possible. This yields / 1  / 1 / . We can define an -dependent cost function         /21     1        which, when minimized, yields      / 1      1 stating that the optimal under this cost function is just the mean of the difference of the Note that   remaining terms.        '<  ' (     (   ( '< and       '<  '  (     (   ( '9 . Several additional assumptions lead to Equation 1. First, let us assume that the prior probabilities are equal  (so   '<  !  ) and that the estimated and true priors are negligibly different (i.e., 1 1  ). Secondly, if we assume that     , this implies that ( ( (   '<:;     >'<=:  '   (  '   under equal priors. While KL-divergence is not sym- ( which means that  >'<  metric in general, we can see that if this holds (or is approximately true for a given problem)  *-,   exactly then the remaining correction is   yielding in Equation 1. To gain further insight, we can examine * , the case when*  the,  likelihoods are Gaussian univariate distributions, with means   and variances   . In this case, *.,  *    ,     , * /   ,  *     ! ,   *# "$   *-, * (2) , It is easy to see that for    the value of is zero for any   . By computing the @ *-, *., * derivative %'&)(+* we can show that is monotonically increasing with  . Hence, is * %', *. ( -  positive iff  2 , , and therefore it penalizes the distribution (class) with higher variance. Note that we have dropped the / argument for notational simplicity. *-, VOCAB SIZE 75 150 300 600 )+*-, WER 2.33584 3.31072 5.22513 7.39268 WER 1.91561 2.89833 4.51365 6.18517 *., Table 1: Word error rates (WER) for likelihood ratio and augmented likelihood ratio )+*-, based classification for various numbers of classes (VOCAB SIZE). Similar relations hold for multivariate Gaussians with means   *-,         * ,  , *   /  *    ,  *  ,      ,   and variances *    *    ,   *   , . (3) The above is zero when the two covariance ( matrices are ( equal. This implies that for Gaussians with equal covariance matrices,  $ '<  ' $  and our correction term is optimal. This is the same analysis (LDA). Moreover, * as the , condition for Fisher?s linear*-discriminant   , we have that ,  for  and *-,  for  in the case with *. , which again implies that penalizes the class that has larger covariance.   2.1 Results *-,  *.,  1   ) on a medium vocabulary speech We tried this method (assuming that 1 recognition task. In our case the likelihood functions     '< are hidden Markov model (HMM) scores3 . The task we chose is NYNEX PHONEBOOK[4], an isolated word speech corpus. Details of the experimental setup, training/test sets, and model topologies, are described in [1]4 . *-, In general, there are a number of ways to compute . These include 1) analytically, using estimated model parameters (possible, for example, with Gaussian densities), 2) computing the KL-divergences on training data using a law-of-large-numbers-like average of likelihood ratios and using training-data estimated model parameters, 3) doing the same as 2 but *., using test data where hypothesized answers come from a first pass -based classification, and 4) Monte-Carlo methods where again the same procedure as 2 is used, but the data is sampled from the training-data estimated distributions. For HMMs, method 1 above is not possible. Also, the data set we used (PHONEBOOK) uses different classes for the training and test sets. In other words, the training and test vocabularies are different. During training, phone models are constructed that are pieced together for the test vocabularies. Therefore, method 2 above is also not possible for this data. Either method 3 or 4 can be used in our case, and we used method 3 in all our experiments. Of course, using the true test labels in method 3 would be the ideal measure of the degree of confusion between models, but these are of course not available (see Figure 2, however, showing the results of a cheating  *-, experiment). Therefore, we use the hypothesized labels . from a first stage to compute The procedure thus is as follows: 1) obtain     '< using maximum likelihood EM training, *-, 2) classify the test set using only and record the error  *., rate, 3) using the hypothesized class labels (answers with errors) to step 2, compute , 4) re-classify the test set using  *-,  *., ) *-, ) *-, (  and record the new error rate. is used if either one of  '#  the score  3 Using 4 state per phone, 12 Gaussian mixtures per state HMMs, totaling 200k free model parameters for the system. 4 Note, however, that error results here are reported on the development set, i.e., PHONEBOOK lists a,b,c,d o,y    VOCAB 75 150 300 600 *-, 2.33584 3.31072 5.22513 7.39268 RAND1 1.87198 2.88505 4.41428 6.15828 RAND500 1.82047 2.71881 4.34608 6.13085 RAND1000 1.91467 2.72809 4.28930 5.91440 WORLD CUP 2.12777 2.79516 3.81583 5.93883 Table 2: The WER under different tournament strategies ( or  $ '< is below a threshold (i.e., when a likely confusion exists), otherwise for classification.  *-, is used Table 1 shows the result of this experiment. The first column shows the vocabulary size of the system (identical to the*-, number of classes)5 . The second column shows the word ) *., error rate (WER) using just , and the third column shows WER using . As can be seen, the WER decreases significantly with this approach. Note also that no additional free parameters are used to obtain these improvements. 3 Playing Games *-, ) *., ( We may view either or as providing a score of class ' over ? when positive, ( class ' wins, and when negative, class wins. In general, the classification procedure may be viewed as a tournament-style game, where for a given sample  , different classes correspond to different players. Players pair together and play each other, and the winner goes on to play another match with a different player. The strategy leading to table 1 required a particular class presentation order ? in that case the order was just the numeric ordering of the arbitrarily assigned integer classes (corresponding to words in this case). *-, alone is used, the order of the comparisons do not matter, leading to Of course when a transitive game [5] (the order of player pairings do not change the final winner). The ) *-, quantity , however, is not guaranteed to be transitive, and when used in a tournament it results in what is called an intransitive game[5]. This means, for example, that might win over  who might win over  who then might win over . Games may be depicted as directed graphs, where an edge between two players point towards the winner. In an intransitive game, the graph contains directed cycles. There has been very little research on intransitive game strategies ? there are in fact a number of philosophical issues relating to if such games are valid or truly exist. Nevertheless, we derived a number of tournament strategies for playing such intransitive games and evaluated their performance in the following. Broadly, there are two tournament types that we considered. Given a particular ordering of the classes     "!#!$!$= %  , we define a sequential tournament when   plays   , the winner plays  , the winner plays  and so on. We also define a tree-based tournament when   plays  ,   plays   , and so on. The tree-based tournament is then applied recursively on the resulting  : winners until a final winner is found. Based on the above, we investigated several intransitive game playing strategies. For RAND1, we just choose a single random tournament order in a sequential tournament. For RAND500, we run 500 sequential tournaments, each one with a different random order. The ultimate winner is taken to be the player who wins the most tournaments. The third strategy plays 1000 rather than 500 tournaments. The final strategy is inspired by worldcup soccer tournaments: given a randomly generated permutation, the class sequence is 5 The 75-word case is an average result of 8 experiments, the 150-word case is an average of 4 cases, and the 300-word case is an average of 2 cases. There are 7291 separate test samples in the 600-word case, and on average about 911 samples per 75-word test case. vocabulary 75 150 300 600  1.0047 1.0061 1.0241 1.0319 var 0.0071 0.0126 0.0551 0.0770  max 2.7662 3.6539 4.0918 5.0460 1.0285 1.0118 1.0170 1.0533 var 0.0759 0.0263 0.0380 0.1482 max 3.8230 3.8724 3.9072 5.5796 Table 3: The statistics of winners. Columns 2-4: 500 random tournaments, Columns 5-7: 1000 random tournaments. separated into 8 groups. We pick the winner of each group using a sequential tournament (the ?regionals?). Then a tree-based tournament is used on the group winners. 60 70 50 60 probability of error (%) probability of error (%) Table 1 compares these different strategies. As can be seen, the results get slightly better (particularly with a larger number of classes) as the number of tournaments increases. Finally, the single word cup strategy does surprisingly well for the larger class sizes. Note that the improvements are statistically significant over the baseline (0.002 using a difference of proportions significance test) and the improvements are more dramatic for increasing vocabulary size. Furthermore, the it appears that the larger vocabulary sizes benefit more from the larger number (1000 rather than 500) of random tournaments. 40 30 20 10 0 50 40 30 20 10 1 2 3 4 length of cycle 5 6 0 0 1 2 3 4 number of cycles detected 5 Figure 1: 75-word vocabulary case. Left: probability of error given that there exists a cycle of at least the given length (a cycle length of one means no cycle found). Right:probability of error given that at least the given number of cycles exist. 3.1 Empirical Analysis In order to better understand our results, this section analyzes the 500 and 1000 random tournament strategies described above. Each set of random tournaments produces a set  of winners which may be described by a histogram. The entropy of that histogram describes its spread, and the number of typical winners is approximately . This is of  ), variance, course relative to each sample  so we may look at the average ( and maximum of this number (the minimum is 1.0 in every case). This is given in Table 3 for the 500 and 1000 cases.  is approximately 1  The table indicates that there is typically only one winner since  and the variances are small. This shows further that the winner is typically not in a cycle, as the existence of a directed cycle in the tournament graph would probably lead to different winners for each random tournament. The relationship between properties of cycles and WER is explored below. When the tournament is intransitive (and therefore the graph possess a cycle), our second analyses shows that the probability of error tends to increase. This is shown in Figure 1 showing that the error probability increases both as the detected cycle length and the num- vocabulary 75 150 300 600 *., 2.33584 3.31072 5.22513 7.39268 skip WER 1.90237 2.76814 4.46296 6.50117 #cycles(%) 13.89 19.6625 22.38 31.96 break WER 1.90223 2.67814 4.46296 6.50117 #cycles(%) 9.34 16.83 21.34 31.53 Table 4: WER results using two strategies (skip and break) thatutilize information about *.,    cycles in the tournament graphs, compared to baseline . The and columns show the number of cycles detected relative to the number of samples in each case. ber of detected cycles increases. 6 This property suggests that the existence of intransitivity could be used as a confidence measure, or could be used to try to reduce errors. As an attempt at the latter, we evaluated two very simple heuristics that try to eliminate cycles as detected during classification. In the first method (skip), we run a sequential tournament (using a random class ordering) until either a clear winner is found (a transitive game), or a cycle is detected. If a cycle is detected, we select two players not in the cycle, effectively jumping out of the cycle, and continue playing until the end of the class ordering. If winner*-, cannot be determined (because there are too few players remaining), we backoff and use to select the winner. In a second method (break), if a cycle is detected, we eliminate the class having the smallest likelihood from that cycle, and then continue playing as before. Neither method detects all the cycles in the graph (their number can be exponentially large). As can be seen, the WER results still provide significant improvements over the baseline, but are no better than earlier results. Because the tournament strategy is coupled with cycle detection, the cycles detected are different in each case (the second method detecting fewer cycles presumably because the eliminated class is in multiple cycles). In any case, it is apparent that further work is needed to investigate the relationship between the existence and properties of cycles and methods to utilize this information. 4 Iterative Determination of KL-divergence In all of our experiments so far, KL-divergence is calculated according to the initial hypothesized answers. We would expect that using the true answers to determine the KLdivergence would improve our results further. The top horizontal lines in Figure 2 shows the original baseline results, and the bottom lines show the results using the true answers (a cheating experiment) to determine the KL-divergence. As can be seen, the improvement is *-, significant thereby confirming that using can significantly improve classification performance. Note also that the relative improvement stays about constant with increasing vocabulary size. This further indicates that an iterative  strategy for determining KL-divergence might fur*-, ther improve our results. In this case, is used to determine the answers to compute the ) *- ,   first set of KL-divergences used in . This is then used to compute a new set of an-  ) *-,  swers which then is used to compute a new scores and so on. The remaining plots in Figure 2 show the results of this strategy for the 500 and 1000 random trials case (i.e., the answers used to compute the KL-divergences in each case are obtained from the previous set of random tournaments using the histogram peak procedure described earlier). Rather surprisingly, the results show that iterating in this fashion does not influence the results in 6 Note that this shows a lower bound on the number of cycles detected. This is saying that if we find, for example, four or more cycles then the chance of error is high. 75 classes 150 classes 3.5 word error rate (%) word error rate (%) 2.5 2 1.5 0 2 4 6 8 number of iterations 3 2.5 2 10 0 2 300 classes 10 600 classes 5.4 7.5 word error rate (%) 5.2 word error rate (%) 4 6 8 number of iterations 5 4.8 4.6 4.4 baseline cheating 500 trials 1000 trials 7 6.5 6 4.2 4 0 2 4 6 8 number of iterations 10 5.5 0 2 4 6 8 number of iterations 10 Figure 2: Baseline using likelihood ratio (top lines), cheating results using correct answers for KL-divergence (bottom lines), and the iterative determination of KL-distance using hypothesized answers from previous iteration (middle lines). any appreciable way ? the WERs seem to decrease only slightly from their initial drop. It is the case, however, that as the number of random tournaments increases, the results become closer to the ideal as the vocabulary size increases. We are currently studying further such iterative procedures for recomputing the KL-divergences. 5 Discussion and Conclusion We have introduced a correction term to the likelihood ratio classification method that is justified by the difference between the *-, estimated and true class conditional probabilities   8 '<>     '< . The correction term is an estimate of the classification bias that would *-, makes the class comoptimally compensate for these differences. The presence of parisons intransitive and we*., introduce several tournament-like strategies to compensate. While the introduction of consistently improves the classification results, further improvements are obtained by the selection of the comparison strategy. Further details and results of our methods will appear in forthcoming publications and technical reports. References [1] J. Bilmes. Natural Statistic Models for Automatic Speech Recognition. PhD thesis, U.C. Berkeley, Dept. of EECS, CS Division, 1999. [2] T.M. Cover and J.A. Thomas. Elements of Information Theory. Wiley, 1991. [3] R.O. Duda, P.E. Hart, and D.G. Stork. Pattern Classification. John Wiley and Sons, Inc., 2000. [4] J. Pitrelli, C. Fong, S.H. Wong, J.R. Spitz, and H.C. Lueng. PhoneBook: A phonetically-rich isolated-word telephone-speech database. In Proc. IEEE Intl. Conf. on Acoustics, Speech, and Signal Processing, 1995. [5] P.D. Straffin. Game Theory and Strategy. The Mathematical ASsociation of America, 1993.
2084 |@word trial:3 middle:1 proportion:1 duda:1 tried:1 covariance:3 pick:1 dramatic:1 thereby:2 recursively:1 reduction:1 initial:2 contains:1 score:3 comparing:2 must:1 john:1 confirming:1 plot:1 drop:1 discrimination:1 nynex:1 alone:1 fewer:1 steal:2 record:2 num:1 detecting:1 mathematical:1 along:1 constructed:1 become:1 pairing:1 introduce:2 theoretically:1 examine:1 multi:1 vocab:3 inspired:1 detects:1 little:1 increasing:4 becomes:1 moreover:2 underlying:1 medium:2 what:1 argmin:1 minimizes:1 kldivergence:1 berkeley:1 every:1 recasts:1 exactly:1 classifier:1 appear:1 positive:7 before:1 engineering:1 dropped:1 tends:1 meil:1 approximately:3 tournament:34 chose:1 might:4 suggests:1 hmms:2 statistically:1 directed:3 practice:2 procedure:7 empirical:2 significantly:4 confidence:3 word:18 get:1 cannot:1 selection:1 risk:1 influence:1 wong:1 go:1 simplicity:1 correcting:2 rule:1 insight:1 notion:1 analogous:1 resp:2 play:8 modulo:1 us:1 element:1 recognition:3 particularly:1 continues:1 asymmetric:1 database:1 bottom:2 negligibly:1 electrical:1 pieced:1 cycle:32 ordering:5 decrease:2 intuition:2 trained:1 division:1 various:1 america:1 separated:1 monte:1 detected:10 tell:1 choosing:1 apparent:1 heuristic:1 larger:5 otherwise:2 favor:4 statistic:3 final:3 sequence:1 iff:1 intuitive:2 seattle:2 intl:1 produce:1 stating:1 stat:1 measured:1 c:1 skip:3 implies:3 come:1 correct:1 viewing:1 secondly:1 adjusted:1 correction:9 hold:2 considered:1 normal:1 presumably:1 smallest:1 purpose:1 proc:1 label:3 currently:1 vice:1 gaussian:4 modified:1 rather:3 totaling:1 publication:1 derived:1 improvement:10 notational:1 fur:1 likelihood:25 indicates:3 consistently:1 baseline:6 dependent:2 typically:2 eliminate:2 hidden:2 relation:1 interested:1 provably:1 issue:2 classification:19 overall:1 development:1 equal:4 having:1 washington:4 eliminated:1 identical:1 look:1 future:1 minimized:1 report:2 inherent:1 few:1 randomly:1 divergence:15 individual:1 argmax:1 attempt:2 detection:1 investigate:3 intransitive:11 adjust:1 mixture:1 truly:1 yielding:2 wers:1 edge:1 closer:1 jumping:1 tree:3 penalizes:2 re:1 isolated:2 instance:1 classify:2 column:6 earlier:2 recomputing:1 phonebook:4 cover:1 cost:2 examining:1 too:1 optimally:1 reported:1 answer:9 fong:1 eec:1 density:3 explores:1 peak:1 stay:1 together:2 again:2 thesis:1 choose:2 conf:1 usable:1 style:3 derivative:1 leading:2 suggesting:1 matter:2 inc:1 view:1 break:3 try:2 analyze:1 doing:1 phonetically:1 variance:5 who:4 correspond:2 yield:3 carlo:1 bilmes:3 justifiable:1 classified:4 explain:1 evaluates:2 spitz:1 mi:3 gain:2 sampled:1 improves:2 appears:2 strat:1 higher:1 evaluated:2 furthermore:3 just:4 implicit:1 lastly:1 stage:1 until:4 horizontal:1 defines:1 lda:1 hypothesized:5 true:13 hence:1 analytically:1 assigned:1 leibler:2 deal:1 game:19 during:2 whereby:1 soccer:1 theoretic:1 confusion:2 ji:1 stork:1 winner:21 exponentially:1 extend:1 association:1 relating:1 significant:4 versa:1 cup:2 automatic:1 access:1 posterior:4 multivariate:1 phone:2 certain:3 arbitrarily:1 continue:2 scoring:2 seen:6 analyzes:1 additional:3 minimum:1 determine:5 monotonically:1 signal:1 multiple:2 technical:1 match:1 determination:2 compensate:2 hart:1 marina:1 rand1:2 metric:1 histogram:3 iteration:5 receive:2 addition:1 justified:1 posse:1 probably:1 seem:1 odds:2 integer:1 ee:1 presence:1 ideal:2 easy:1 variety:1 forthcoming:1 topology:1 reduce:1 ultimate:1 speech:7 iterating:1 clear:1 backoff:1 exist:2 estimated:11 per:3 broadly:1 express:1 group:3 four:1 threshold:1 nevertheless:1 changing:1 neither:1 utilize:2 graph:6 run:2 wer:12 saying:1 decision:2 bound:1 followed:1 guaranteed:1 gang:1 erroneously:1 aspect:1 argument:1 department:2 according:1 describes:1 slightly:2 em:1 son:1 taken:1 equation:2 discus:1 needed:1 end:1 studying:1 available:1 gaussians:3 observe:1 appropriate:1 existence:3 original:1 thomas:1 top:2 remaining:4 include:1 already:1 quantity:3 added:1 strategy:24 win:6 distance:2 separate:1 hmm:1 discriminant:1 assuming:1 length:4 relationship:2 gji:1 ratio:18 providing:1 equivalently:1 setup:1 negative:3 motivates:1 unknown:1 markov:2 extended:1 introduced:1 pair:3 cheating:4 specified:1 kl:15 required:1 philosophical:1 intransitivity:3 acoustic:1 ther:1 proceeds:1 below:2 pattern:2 including:1 max:2 explanation:1 confusability:1 event:1 natural:1 improve:5 concludes:1 transitive:4 coupled:1 prior:5 determining:1 relative:3 law:1 loss:3 expect:1 permutation:2 var:2 degree:1 sufficient:2 playing:8 course:5 surprisingly:2 free:3 sym:1 bias:4 understand:2 ber:1 benefit:1 calculated:1 vocabulary:12 evaluating:1 world:1 numeric:1 valid:1 rich:1 far:1 kullback:2 corpus:1 iterative:6 why:1 table:9 improving:1 investigated:1 significance:1 spread:1 augmented:1 fashion:1 aid:1 wiley:2 mmp:1 third:2 showing:2 list:1 explored:1 exists:2 sequential:5 effectively:1 phd:1 egy:1 entropy:1 depicted:1 simply:1 likely:4 univariate:1 chance:1 conditional:3 viewed:2 presentation:1 towards:1 appreciable:1 jeff:1 fisher:1 change:1 typical:1 determined:1 telephone:1 called:2 pas:1 experimental:2 player:10 indicating:1 select:2 latter:1 dept:1 tested:1
1,190
2,085
Multiplicative Updates for Classification by Mixture Models  Lawrence K. Saul and Daniel D. Lee Department of Computer and Information Science  Department of Electrical Engineering University of Pennsylvania, Philadelphia, PA 19104 Abstract We investigate a learning algorithm for the classification of nonnegative data by mixture models. Multiplicative update rules are derived that directly optimize the performance of these models as classifiers. The update rules have a simple closed form and an intuitive appeal. Our algorithm retains the main virtues of the Expectation-Maximization (EM) algorithm?its guarantee of monotonic improvement, and its absence of tuning parameters?with the added advantage of optimizing a discriminative objective function. The algorithm reduces as a special case to the method of generalized iterative scaling for log-linear models. The learning rate of the algorithm is controlled by the sparseness of the training data. We use the method of nonnegative matrix factorization (NMF) to discover sparse distributed representations of the data. This form of feature selection greatly accelerates learning and makes the algorithm practical on large problems. Experiments show that discriminatively trained mixture models lead to much better classification than comparably sized models trained by EM. 1 Introduction Mixture models[11] have been widely applied to problems in classification. In these problems, one must learn a decision rule mapping feature vectors (   ) to class labels (  ) given labeled examples. Mixture models are typically used to parameterize class-conditional dis  , and then to compute posterior probabilities,      , from Bayes rule. tributions,    Parameter estimation in these models is handled by an Expectation-Maximization (EM) algorithm[3], procedure that monotonically increases the joint log likelihood, a learning        , summed over training examples (indexed by  ). A virtue of this algorithm is that it does not require the setting of learning rates or other tuning parameters. A weakness of the above approach is that the model parameters are optimized by maximum likelihood estimation, as opposed to a discriminative criterion more closely related to classification error[14]. In this paper, we derive multiplicative update rules for the parameters   models that directly maximize the discriminative objective function,   of mixture     . This objective function measures the conditional log likelihood that the training examples are correctly classified. Our update rules retain the main virtues of the EM algorithm?its guarantee of monotonic improvement, and its absence of tuning parameters?with the added advantage of optimizing a discriminative cost function. They also have a simple closed form and appealing intuition. The proof of convergence combines ideas from the EM algorithm[3] and methods for generalized and improved iterative scaling[2, 4]. The approach in this paper is limited to the classification of nonnegative data, since from the constraint of nonnegativity emerges an especially simple learning algorithm. This limitation, though, is not too severe. An abundance of interesting data occurs naturally in this form: for example, the pixel intensities of images, the power spectra of speech, and the word-document counts of text. Real-valued data can also be coerced into this form by addition or exponentiation. Thus we believe the algorithm has broad applicability. 2 Mixture models as generative models Mixture models are typically used  as generative models to parameterize probability distributions over feature vectors   . Different mixture models are used to model different classes of data. The parameterized distributions take the form:         (1)   rows     the where of the nonnegative weight matrix  are constrained to sum to unity,    ,  and the basis functions   are properly normalized distributions, such  that   for all  . The model can be interpreted as the latent variable model,                  (2)     is  used where the discrete latent variable  indicates which mixture component to gen  erate the observed variable  . In this setting, one identifies        and          . The basis functions, usually chosen from the exponential family,      define ?bumps? of high probability in the feature space. A popular choice is the multivariate Gaussian distribution:      !       3   40"6 5 #    3   7   (3)  " $#&%('*),+.-0/21 1 1 3 " with means  and covariance matrices . Gaussian distributions are extremely versatile, but not always the most appropriate. For sparse nonnegative data, a more natural choice is the exponential distribution:    98;:<  :>= 5?@BA CA    <  with parameter vectors . Here, the value of 3  <  indexes the elements of parameters of these basis functions must be estimated from data. (4) and   . The Generative models can be viewed as a prototype method for classification, with the parameters of each mixture component defining a particular basin of attraction in the feature space. Intuitively, patterns are labeled by the most similar prototype, chosen from among all possible classes. Formally, unlabeled examples are classified by computing posterior probabilities from Bayes? rule,       EDGF         H    IH  (5) where    denote the prior probabilities of each class. Examples are classified by the label with the highest posterior probability. An Expectation-Maximization (EM) algorithm can be used to estimate the parameters of mixture models. The EM algorithm optimizes the joint log likelihood,                 (6) summed over training examples. If basis functions are not shared across different classes,   can be done independently for each class label  . then the parameter estimation for    This has the tremendous advantage of decomposing the original learning problem into several smaller problems. Moreover, for many types of basis functions, the EM updates have a simple closed form and are guaranteed to improve the joint log likelihood at each iteration. These properties account for the widespread use of mixture models as generative models. 3 Mixture models as discriminative models Mixture models can also be viewed as purely discriminative models. In this view, their purpose is simply to provide a particular way of parameterizing the posterior distribution,     . In this paper, we study posterior distributions of the form:                    (7)  .a valid The right hand side  of this equation defines posterior distribution provided that the    are nonnegative. Note that for this intermixture weights and basis functions  pretation, the mixture weights and basis functions do not need to satisfy the more stringent normalization constraints of generative models. We will deliberately exploit this freedom, an idea that distinguishes our approach from previous work on discriminatively trained mixture models[6] and hidden Markov models[5, 12]. In particular, the unnormalized basis functions we use are able to parameterize ?saddles? and ?valleys? in the feature space, as well as the ?bumps? of normalized basis functions. This makes them more expressive than their generative counterparts: examples can not only be attracted to prototypes, but also repelled by opposites.    The posterior distributions in eq. (7) must be further specified by parameterizing the basis   as a function of   . We study basis functions of the form functions      = @   (8)  denotes a real-valued vector and   where denotes a nonnegative and possibly ?expanded? representation[14] of the original feature vector. The exponential form in eq. (8) allows us to recover certain generative models as a special case. For example, consider the multivariate Gaussian distribution in eq. (3). By defining the ?quadratically expanded? feature vector:                       (9)   # ' # # # ' 5#     (8) by choosing the parameter  vectors we can equate the basis functions in eqs. (3) and  to act on  in the same way that the means 3  and covariance matrices " act on   . The exponential distributions in eq. (4) can be recovered in a similar way. Such generative models provide a cheap way to initialize discriminative models for further training. 4 Learning algorithm Our learning algorithm directly optimizes the performance of the models in eq. (7) as classifiers. The objective function we use for discriminative training is the conditional log likelihood,      (10)      9    summed over training examples. Let denote the binary matrix whose  th element denotes whether the  th training example belongs to Then we can write the  the th class., where: objective function as the difference of two terms,   5          1   5 =    @  E  = @    (11) (12) 5 The competition between these terms give rise to a scenario of contrastive learning. It is the subtracted term, , which distinguishes the conditional log likelihood optimized by discriminative training from the joint log likelihood optimized by EM. Our learning algorithm works by alternately updating the mixture weights and the basis function parameters. Here we simply present the update rules for these parameters; a derivation and proof of convergence are given in the appendix. It is easiest to write the basis function updates in terms of the nonnegative parameters . The updates then take the simple multiplicative form:   = @A     / = @A / = @BA            5  7             +    :   5 : 7 where   : : :  (13) (14) It is straightforward to compute the gradients in these ratios and show that they are always nonnegative. (This is a consequence of the nonnegativity constraint on the feature vectors:    for all examples  and feature components .) Thus, the nonnegativity constraints on the mixture weights and basis functions are enforced by these multiplicative udpates. 3 The updates have a simple intuition[9] based on balancing opposing terms in the gradient of the conditional log likelihood. In particular, note that the fixed points of this  at stationary update    rule occur     points of the conditional log  likelihood?that  is, where    and , or equivalently, where        and . The learning rate is controlled by the ratios  of these gradients and?additionally, for the basis function updates?by the exponent  , which measures the sparseness of the training data. The value of  is the maximum sum of features that occurs in the training data. Thus, sparse feature vectors leads to faster learning, a crucial point to which we will return shortly.    :  5  : :        5 It is worth comparing these multiplicative updates to others in the literature. Jebara and Pentland[6] derived similar updates for mixture weights, but without emphasizing the special form of eq. (13). Others have investigated multiplicative updates by the method of exponentiated gradients (EG)[7]. Our updates do not have the same form as EG updates: in particular, note that the gradients in eqs. (13?14) are not exponentiated. If we use one basis function per class and an identity matrix for the mixture weights, then the updates reduce to the method of generalized iterative scaling[2] for logistic or multinomial regression (also known as maximum entropy modeling). More generally, though, our multiplicative updates can be used to train much more powerful classifiers based on mixture models. 5 Feature selection As previously mentioned, the learning rate for the basis function parameters is controlled by the sparseness of the training data. If this data is not intrinsically sparse, then the multiplicative upates in eqs. (13?14) can be impractically slow (just as the method of iterative pixel image NMF basis vectors 01-10 11-20 21-30 31-40 NMF feature vector 10 41-50 51-60 5 61-70 71-80 0 20 40 60 80 Figure 1: Left: nonnegative basis vectors for handwritten digits discovered by NMF. Right: sparse feature vector for a handwritten ?2?. The basis vectors are ordered by their contribution to this image. scaling). In this case, it is important to discover sparse distributed representations of the data that encode the same information. On large problems, such representations can accelerate learning by several orders of magnitude. The search for sparse distributed representations can be viewed as a form of feature selection. We have observed that suitably sparse representations can be discovered by the method of nonnegative matrix factorization (NMF)[8]. Let the raw nonnegative (and posmatrix , where is its raw dimensibly nonsparse) data be represented by the sionality and  is the number of training examples. Algorithms a factor for NMF yield ization , where is a nonnegative marix and is a nonnegative matrix. In this  factorization, the columns of are interpreted as basis vectors, and the columns of as coefficients (or new feature vectors). These coefficients are typically very sparse, because the nonnegative basis vectors can only be added in a constructive way to approximate the original data.            The effectiveness of NMF is best illustrated by example. We used the method to discover sparse distributed representations of the MNIST data set of handwritten digits[10]. The data set has  60000 training and 10000 test examples that were deslanted and cropped to      training data was therefore represented by form grayscale pixel  images. The raw  of Fig. 1 shows the a matrix, with and . The  left plot basis vectors discovered by NMF, each plotted as a image. Most of these basis vectors resemble strokes, only a fraction of which are needed to reconstruct any particular image in the training set. For example, only about twenty basis vectors make an appreciable contribution to the handwritten ?2? shown in the right plot of Fig. 1. The method of NMF thus succeeds in discovering a highly sparse representation of the original images.     6          Results Models were evaluated on the problem of recognizing handwritten digits from the MNIST data set. From the grayscale pixel images, we generated two vectors: one  sets ; ofthefeature by NMF, with nonnegative features and dimensionality other, by principal      . These components analysis (PCA), with real-valued features and dimensionality  reduced dimensionality feature vectors were used for both training and testing. Baseline mixture models for classification were trained by EM algorithms. Gaussian mixture models with diagonal covariance marices were trained on the PCA features, while exponential mixture models (as in eq. (4)) were trained on the NMF features. The mixture models were trained for up to 64 iterations of EM, which was sufficient to ensure a high degree of convergence. Seven baseline classifiers were trained set, with   on each  feature   different numbers of mixture components per digit ( ). The error rates of these models, indicated by EM-PCA40 and EM-NMF80, are shown in Table 1. Half as many PCA features were used as NMF features so as to equalize the number of fitted parameters in different basis functions.              Mixture models on the NMF features were also trained discriminatively by the multiplicative updates in Models with varying numbers of mixture components per  eqs.   (13?14).   ) were trained by 1000 iterations of these updates. Again, this was digit ( sufficient to ensure a high degree of convergence; there was no effort at early stopping.    for randomly selected feaThe models were initialized by setting and  ture vectors. The results of these experiments, indicated by DT-NMF80, are also shown in Table 1. The results show that the discriminatively trained models classify much better than comparably sized models trained by EM. The ability to learn more compact classifiers appears to be the major advantage of discriminative training. A slight disadvantage is that the resulting classifiers are more susceptible to overtraining. model        EM-PCA40   1 2 4 8 16 32 64 10.2 8.5 6.8 5.3 4.0 3.1 1.9 10.1 8.3 6.4 5.1 4.4 3.6 3.1   EM-NMF80    15.7 12.3 9.3 7.8 6.2 5.0 3.9  DT-NMF80   14.7 10.7 8.2 7.0 5.7 5.1 4.2 5.5 4.0 2.8 1.7 1.0 5.8 4.4 3.5 3.2 3.4 Table 1: Classification error rates (%) on the training set ( and the test set (  ) for mixture models with different numbers of mixture components per digit (  the same number of fitted parameters. ). Models in the same row have roughly   It is instructive to compare our results to other benchmarks on this data set[10]. Without making use of prior knowledge, better error rates on the test been obtained by sup  set have  ), and fully connected port vector machines ( neighbor (    ), k-nearest  multilayer neural networks (   ). These results, however, either required storing large numbers of training examples or training significantly larger models. For example, the nearest neighbor and support vector classifiers required storing tens of thousands of training examples  (or support vectors), while the neural network had over 120,000 By  weights.  contrast, the discriminatively trained mixture model (with   ) has less than 6500 iteratively adjusted parameters, and most of its memory footprint is devoted to preprocessing by NMF.          We conclude by describing the problems best suited to the mixture models in this paper. These are problems with many classes, large amounts of data, and little prior knowledge of symmetries or invariances. Support vector machines and nearest neighbor algorithms do not scale well to this regime, and it remains tedious to train large neural networks with unspecified learning rates. By contrast, the compactness of our models and the simplicity of their learning algorithm make them especially attractive. A Proof of convergence In this appendix, we show that the multiplicative updates from section 4 lead to monotonic improvement in the conditional log likelihood. This guarantee of convergence (to a stationary point) is proved by computing a lower bound on the conditional log likelihood for updated estimates of the mixture weights and basis function parameters. We indicate these  updated estimates by and  , and we indicate the resulting values of the conditional  , and . The proof of convergence rests log likelihood and its component terms by , on three simple inequalities applied to .  H  H H H H H5 The first term in the conditional log likelihood can be lower bounded by Jensen?s inequality. The same bound is used here as in the derivation of the EM algorithm[3, 13] for maximum likelihood estimation:     = @F  F =      H          H 9        H @ (15)     The right hand side of this inequality introduces an auxiliary probability distribution    bound holds for arbitrary distributions, provided   The for each example in the training  set. they are properly normalized:    for all  .  The second term in the conditional log likelihood occurs with a minus sign, so for this term we require an upper bound. The same bounds can be used here as in derivations  ofiterative  scaling[1, 2, 4, 13]. Note that the logarithm function is upper bounded by:   for all . We can therefore write:     = @F       H  =  1   (16) H5 1 5        To further bound =the  hand side of eq. (16), we make the following  observation: though @F right  H : ,  H the exponentials are convex functions of the parameter vector with elements F =  =  F @ @BA , they are concave functions of the ?warped? parameter vector with elements where  is defined by eq. (14).  (The validity of this observation hinges on the nonnegativity of the = feature =  F  set,=  the F   vectors   .) It follows that for any example in the training @BA , @ @BA is upper bounded by its linearized expansion around exponential given by: . : = F   = F   = @    :  = @BA 1 = @ A "!   =  @@BA #  @ (17) = =  @BA in eq. (17) is the derivative of @  with respect to the The last term in parentheses , computed by the chain rule. Tighter bounds are possible than independent variable eq. (17), but at the expense of more complicated update rules.   Combining the above inequalities with a judicious choice for the auxiliary parameters  , we obtain a proof of convergence for the multiplicative updates in eqs. (13?14). Let: = =             5 #   @   (18)  =  =  #     5    @    5      (19)   Eq. (18) sets the auxiliary parameters  , while eq. (19) defines an analogous distribu 5  = F    H   = @      1   tion for the opposing term in the conditional log likelihood. (This will prove to be a useful notation.) Combining these definitions with eqs. (15?17) and rearranging terms, we obtain the following inequality: H1     H       /        H        H   5       1  1      1        7 = F 5 @BA  @BA 1 # (20)      Both sides of the inequality vanish (yielding an equality) if  H   and  H   . We derive the update rules by maximizing the right hand side of this inequality. Maximizing the right hand side with respect to  H while holding the basis function parameters fixed    yields the update, eq. (13). Likewise, maximizing the right hand side with respect to H    the mixture weights fixed yields the update, eq. (14). Since these choices while holding  H and H lead to positive values on the right hand side of the inequality (except at for H   : :   ! fixed points), it follows that the multiplicative updates in eqs. (13?14) lead to monotonic improvement in the conditional log likelihood. References [1] M. Collins, R. Schapire, and Y. Singer (2000). Logistic regression, adaBoost, and Bregman distances. In Proceedings of the Thirteenth Annual Conference on Computational Learning Theory. [2] J. N. Darroch and D. Ratcliff (1972). Generalized iterative scaling for log-linear models. Annals of Mathematical Statistics 43:1470?1480. [3] A. P. Dempster, N. M. Laird, and D. B. Rubin (1977). Maximum likelihood from incomplete data via the EM algorithm. J. Royal Stat. Soc. B 39: 1?37. [4] S. Della Pietra, V. Della Pietra, and J. Lafferty (1997). Inducing features of random fields. IEEE Transactions on Pattern Analysis and Machine Intelligence 19(4): 380?393. [5] P.S. Gopalakrishnan, D. Kanevsky, A. Ndas and D. Nahamoo (1991). An inequality for rational functions with applications to some statistical estimation problems. IEEE Transactions on Information Theory 37: 107?113. [6] T. Jebara and A. Pentland (1998). Maximum conditional likelihood via bound maximization and the CEM algorithm. In M. Kearns, S. Solla, and D. Cohn (eds.). Advances in Neural Information Processing Systems 11, 494?500. MIT Press: Cambridge, MA. [7] J. Kivinen and M. Warmuth (1997). Additive versus exponentiated gradient updates for linear prediction. Journal of Information and Computation 132: 1?64. [8] D. D. Lee and H. S. Seung (1999). Learning the parts of objects with nonnegative matrix factorization. Nature 401: 788?791. [9] D. D. Lee and H. S. Seung (2000). Algorithms for nonnegative matrix factorization. In T. Leen, T. Dietterich, and V. Tresp (eds.). Advances in Neural Information Processing Systems 13. MIT Press: Cambridge, MA. [10] Y.LeCun, L. Jackel, L.Bottou, A.Brunot, C.Cortes, J. Denker, H.Drucker, I.Guyon, U. Muller, E.Sackinger, P.Simard, and V.Vapnik (1995). A comparison of learning algorithms for handwritten digit recognition. In F.Fogelman and P.Gallinari (eds.). International Conference on Artificial Neural Networks, 1995, Paris: 53?60. [11] G. McLachlan and K. Basford (1988). Mixture Models: Inference and Applications to Clustering. Marcel Dekker. [12] Y. Normandin (1991). Hidden Markov Models, Maximum Mutual Information Estimation and the Speech Recognition Problem. Ph.D. Thesis, McGill University, Montreal. [13] J. A. O?Sullivan (1998). Alternating minimization algorithms: from Blahut-Arimoto to Expectation-Maximization. In A. Vardy (ed.). Codes, Curves, and Signals: Common Threads in Communications. Kluwer: Norwell, MA. [14] V. Vapnik (1999). The Nature of Statistical Learning Theory. Springer Verlag.
2085 |@word erate:1 suitably:1 tedious:1 dekker:1 linearized:1 covariance:3 contrastive:1 minus:1 versatile:1 daniel:1 document:1 recovered:1 comparing:1 must:3 attracted:1 additive:1 cheap:1 plot:2 update:29 stationary:2 generative:8 discovering:1 half:1 selected:1 intelligence:1 warmuth:1 mathematical:1 prove:1 combine:1 roughly:1 little:1 provided:2 discover:3 moreover:1 bounded:3 notation:1 easiest:1 interpreted:2 unspecified:1 guarantee:3 act:2 concave:1 classifier:7 gallinari:1 positive:1 engineering:1 consequence:1 factorization:5 limited:1 practical:1 lecun:1 testing:1 della:2 footprint:1 sullivan:1 digit:7 procedure:1 significantly:1 word:1 unlabeled:1 selection:3 valley:1 optimize:1 maximizing:3 straightforward:1 independently:1 convex:1 simplicity:1 rule:12 attraction:1 parameterizing:2 analogous:1 updated:2 annals:1 mcgill:1 pa:1 element:4 recognition:2 updating:1 tributions:1 labeled:2 observed:2 electrical:1 parameterize:3 thousand:1 connected:1 solla:1 highest:1 mentioned:1 intuition:2 equalize:1 dempster:1 seung:2 trained:13 purely:1 basis:29 accelerate:1 joint:4 represented:2 derivation:3 train:2 artificial:1 choosing:1 whose:1 widely:1 valued:3 larger:1 reconstruct:1 ability:1 statistic:1 laird:1 advantage:4 combining:2 gen:1 intuitive:1 inducing:1 competition:1 convergence:8 object:1 derive:2 montreal:1 stat:1 nearest:3 eq:22 soc:1 auxiliary:3 resemble:1 indicate:2 marcel:1 closely:1 stringent:1 require:2 tighter:1 adjusted:1 hold:1 around:1 lawrence:1 mapping:1 bump:2 major:1 early:1 purpose:1 estimation:6 label:3 jackel:1 mclachlan:1 minimization:1 mit:2 gaussian:4 always:2 varying:1 encode:1 derived:2 improvement:4 properly:2 likelihood:20 indicates:1 ratcliff:1 greatly:1 contrast:2 baseline:2 inference:1 stopping:1 typically:3 compactness:1 hidden:2 pixel:4 fogelman:1 classification:9 among:1 exponent:1 constrained:1 special:3 summed:3 initialize:1 mutual:1 field:1 broad:1 others:2 distinguishes:2 randomly:1 pietra:2 blahut:1 opposing:2 freedom:1 investigate:1 highly:1 severe:1 weakness:1 introduces:1 mixture:36 yielding:1 devoted:1 chain:1 norwell:1 bregman:1 indexed:1 incomplete:1 logarithm:1 initialized:1 plotted:1 fitted:2 column:2 modeling:1 classify:1 disadvantage:1 retains:1 maximization:5 cost:1 applicability:1 recognizing:1 too:1 international:1 retain:1 lee:3 deslanted:1 again:1 thesis:1 opposed:1 possibly:1 warped:1 derivative:1 simard:1 return:1 account:1 coefficient:2 satisfy:1 multiplicative:13 view:1 tion:1 closed:3 h1:1 sup:1 bayes:2 recover:1 complicated:1 contribution:2 equate:1 likewise:1 yield:3 handwritten:6 raw:3 comparably:2 worth:1 classified:3 stroke:1 overtraining:1 ed:4 definition:1 naturally:1 proof:5 basford:1 rational:1 proved:1 popular:1 intrinsically:1 knowledge:2 emerges:1 dimensionality:3 appears:1 dt:2 adaboost:1 improved:1 leen:1 done:1 though:3 evaluated:1 just:1 hand:7 expressive:1 cohn:1 sackinger:1 widespread:1 defines:2 logistic:2 indicated:2 believe:1 dietterich:1 validity:1 normalized:3 deliberately:1 counterpart:1 equality:1 ization:1 alternating:1 iteratively:1 illustrated:1 eg:2 attractive:1 nonsparse:1 unnormalized:1 criterion:1 generalized:4 image:8 common:1 multinomial:1 arimoto:1 slight:1 kluwer:1 cambridge:2 tuning:3 had:1 posterior:7 multivariate:2 optimizing:2 optimizes:2 belongs:1 scenario:1 certain:1 verlag:1 inequality:9 binary:1 muller:1 maximize:1 monotonically:1 signal:1 reduces:1 faster:1 controlled:3 parenthesis:1 prediction:1 coerced:1 regression:2 multilayer:1 expectation:4 iteration:3 normalization:1 addition:1 cropped:1 thirteenth:1 crucial:1 rest:1 lafferty:1 effectiveness:1 ture:1 kanevsky:1 pennsylvania:1 opposite:1 reduce:1 idea:2 prototype:3 drucker:1 whether:1 thread:1 handled:1 pca:3 darroch:1 effort:1 speech:2 generally:1 useful:1 amount:1 ten:1 ph:1 reduced:1 schapire:1 sign:1 estimated:1 correctly:1 per:4 discrete:1 write:3 fraction:1 sum:2 enforced:1 exponentiation:1 parameterized:1 powerful:1 family:1 guyon:1 decision:1 appendix:2 scaling:6 accelerates:1 bound:8 guaranteed:1 nonnegative:18 annual:1 nahamoo:1 occur:1 constraint:4 extremely:1 expanded:2 department:2 across:1 smaller:1 em:18 unity:1 appealing:1 making:1 brunot:1 intuitively:1 equation:1 previously:1 remains:1 describing:1 count:1 needed:1 singer:1 decomposing:1 denker:1 appropriate:1 subtracted:1 shortly:1 original:4 denotes:3 clustering:1 ensure:2 hinge:1 exploit:1 especially:2 objective:5 added:3 occurs:3 diagonal:1 gradient:6 distance:1 seven:1 gopalakrishnan:1 code:1 index:1 ratio:2 equivalently:1 susceptible:1 holding:2 expense:1 repelled:1 rise:1 ba:10 twenty:1 upper:3 observation:2 markov:2 benchmark:1 pentland:2 defining:2 communication:1 discovered:3 arbitrary:1 jebara:2 intensity:1 nmf:14 upates:1 required:2 specified:1 paris:1 optimized:3 quadratically:1 tremendous:1 alternately:1 able:1 usually:1 pattern:2 regime:1 royal:1 memory:1 power:1 natural:1 kivinen:1 improve:1 identifies:1 philadelphia:1 tresp:1 text:1 prior:3 literature:1 fully:1 discriminatively:5 interesting:1 limitation:1 versus:1 degree:2 basin:1 sufficient:2 rubin:1 port:1 storing:2 balancing:1 row:2 last:1 dis:1 distribu:1 side:8 exponentiated:3 saul:1 neighbor:3 sparse:11 distributed:4 curve:1 valid:1 preprocessing:1 transaction:2 approximate:1 compact:1 cem:1 conclude:1 discriminative:10 spectrum:1 grayscale:2 search:1 iterative:6 latent:2 table:3 additionally:1 learn:2 nature:2 rearranging:1 symmetry:1 expansion:1 investigated:1 bottou:1 main:2 fig:2 slow:1 nonnegativity:4 exponential:7 vanish:1 abundance:1 emphasizing:1 jensen:1 appeal:1 cortes:1 virtue:3 ih:1 mnist:2 vapnik:2 magnitude:1 sparseness:3 suited:1 entropy:1 simply:2 saddle:1 ordered:1 monotonic:4 springer:1 ma:3 conditional:14 sized:2 viewed:3 identity:1 appreciable:1 shared:1 absence:2 judicious:1 except:1 impractically:1 principal:1 kearns:1 invariance:1 succeeds:1 formally:1 support:3 collins:1 constructive:1 h5:2 instructive:1
1,191
2,086
Estimating the Reliability of leA Projections F. Meinecke l ,2, A. Ziehe l , M. Kawanabe l and K.-R. Miiller l ,2* 1 Fraunhofer FIRST.IDA, Kekuh~str. 7, 12489 Berlin, Germany 2University of Potsdam, Am Neuen Palais 10, 14469 Potsdam, Germany {meinecke,ziehe,nabe,klaus}?first.fhg.de Abstract When applying unsupervised learning techniques like ICA or temporal decorrelation, a key question is whether the discovered projections are reliable. In other words: can we give error bars or can we assess the quality of our separation? We use resampling methods to tackle these questions and show experimentally that our proposed variance estimations are strongly correlated to the separation error. We demonstrate that this reliability estimation can be used to choose the appropriate ICA-model, to enhance significantly the separation performance, and, most important, to mark the components that have a actual physical meaning. Application to 49-channel-data from an magneto encephalography (MEG) experiment underlines the usefulness of our approach. 1 Introduction Blind source separation (BSS) techniques have found wide-spread use in various application domains , e.g. acoustics , telecommunication or biomedical signal processing. (see e.g. [9, 5, 6, 1, 2, 4, 14, 8]). BSS is a statistical technique to reveal unknown source signals when only mixtures of them can be observed. In the following we will only consider linear mixtures; the goal is then to estimate those projection directions, that recover the source signals. Many different BSS algorithms have been proposed, but to our knowledge, so far, no principled attempts have been made to assess the reliability of BSS algorithms, such that error bars are given along with the resulting projection estimates. This lack of error bars or means for selecting between competing models is of course a basic dilemma for most unsupervised learning algorithms. The sources of potential unreliability of unsupervised algorithms are ubiquous , i.e. noise, non-stationarities, small sample size or inadequate modeling (e.g. sources are simply dependent instead of independent). Unsupervised projection techniques like PCA or BSS will always give an answer that is found within their model class, e.g. PCA will supply an orthogonal basis even if the correct modeling might be non-orthogonal. But how can we assess such a miss-specification or a large statistical error? Our approach to this problem is inspired by the large body of statistics literature on ? To whom correspondence should be addressed. resampling methods (see [12] or [7] for references), where algorithms for assessing the stability of the solution have been analyzed e.g. for peA [3]. We propose reliability estimates based on bootstrap resampling. This will enable us to select a good BSS model, in order to improve the separation performance and to find potentially meaningful projection directions. In the following we will give an algorithmic description of the resampling methods, accompanied by some theoretical remarks (section 2) and show excellent experimental results (sections 3 and 4). We conclude with a brief discussion. 2 2.1 Resampling Techniques for BSS The leA Model In blind source separation we assume that at time instant t each component Xi(t) of the observed n-dimensional data vector, x(t) is a linear superposition of m ::::: n statistically independent signals: m Xi(t) = LAijSj(t) j=l (e.g. [8]). The source signals Sj(t) are unknown, as are the coefficients Aij of the mixing matrix A. The goal is therefore to estimate both unknowns from a sample of the x(t), i.e. y(t) = s(t) = Wx(t), where W is called the separating matrix. Since both A and s(t) are unknown, it is impossible to recover the scaling or the order of the columns of the mixing matrix A. All that one can get are the projection directions. The mixing/ demixing process can be described as a change of coordinates. From this point of view the data vector stays the same, but is expressed in different coordinate systems (passive transformation). Let {ed be the canonical basis of the true sources s = 'E eiSi. Analogous, let {fj} be the basis of the estimated leA channels: y = 'E fjYj. Using this, we can define a component-wise separation error Ei as the angle difference between the true direction of the source and the direction of the respective leA channel: Ei = arccos ("e~i: ~ifill) . To calculate this angle difference, remember that component-wise we have Yj 'E WjkAkisi. With Y = s, this leads to: fj = 'E ei(WA)ij1, i.e. fj is the j-th column of (WA) - l. In the following, we will illustrate our approach for two different source separation algorithms (JADE, TDSEP). JADE [4] using higher order statistics is based on the joint diagonalization of matrices obtained from 'parallel slices' of the fourth order cumulant tensor. TDSEP [14] relies on second order statistics only, enforcing temporal decorrelation between channels. 2.2 About Resampling The objective of resampling techniques is to produce surrogate data sets that eventually allow to approximate the 'separation error' by a repeated estimation of the parameters of interest. The underlying mixing should of course be independent of the generation process of the surrogate data and therefore remain invariant under resampling. Bootstrap R esampling The most popular res amp ling methods are the Jackknife and the Bootstrap (see e.g. [12, 7]) The Jackknife produces surrogate data sets by just deleting one datum each time from the original data. There are generalizations of this approach like k-fold cross-validation which delete more than one datum at a time. A more general approach is the Bootstrap. Consider a block of, say, N data points. For obtaining one bootstrap sample, we draw randomly N elements from the original data, i.e. some data points might occur several times, others don't occur at all in the bootstrap sample. This defines a series {at} with each at telling how often the data point x(t) has been drawn. Then, the separating matrix is computed on the full block and repeatedly on each of the N -element bootstrap samples. The variance is computed as the squared average difference between the estimate on the full block and the respective bootstrap unmixings. (These resampling methods have some desirable properties, which make them very attractive; for example, it can be shown that for iid data the bootstrap estimators of the distributions of many commonly used statistics are consistent.) It is straight forward to apply this procedure to BSS algorithms that do not use time structure; however , only a small modification is needed to take time structure into account. For example, the time lagged correlation matrices needed for TDSEP, can be obtained from {ad by 1 Cij(T) = N N 2: at 'Xi(t)Xj(t+T) t= l with L at = N and at E {O, 1, 2, ... }. Other resampling methods Besides the Bootstrap, there are other res amp ling methods like the Jackknife or cross-validation which can be understood as special cases of Bootstrap. We have tried k-fold cross-validation, which yielded very similar results to the ones reported here. 2.3 The Resampling Algorithm After performing BSS, the estimated ICA-projections are used to generate surrogate data by resampling. On the whitened l surrogate data, the source separation algorithm is used again to estimate a rotation that separates this surrogate data. In order to compare different rotation matrices, we use the fact that the matrix representation of the rotation group SO(N) can be parameterized by r5~r5t - r5~r5b , where the matrices Mij are generators of the group with (Mab)ij and the aij are the rotation parameters (angles) of the rotation matrix R. Using this parameterization we can easily compare different N-dimensional rotations by comparing the rotation parameters aij. Since the sources are already separated, the estimated rotation matrices will be in the vicinity of the identity matrix.2 . IThe whitening transformation is defined as x' = Vx with V = E[xxTtl/2. 21t is important to perform the resampling when the sources are already separated, so that the aij are distributed around zero, because SO(N) is a non-Abelian group; that means that in general R(a)R?(3 ) of- R?(3) R(a) . Var(aij) measures the instability of the separation with respect to a rotation in the (i, j)-plane. Since the reliability of a projection is bounded by the maximum angle variance of all rotations that affect this direction, we define the uncertainty of the i-th ICA-Projection as Ui := maxj Var(aij). Let us summarize the resampling algorithm: 1. Estimate the separating matrix W with some ICA algorithm. Calculate the ICA-Projections y = Wx 2. Produce k surrogate data sets from y and whiten these data sets 3. For each surrogate data set: do BSS, producing a set of rotation matrices 4. Calculate variances of rotation parameters (angles) aij 5. For each ICA component calculate the uncertainty Ui = maxVar(aij). J 2.4 Asymptotic Considerations for Resampling Properties of res amp ling methods are typically studied in the limit when the number of bootstrap samples B -+ 00 and the length of signal T -+ 00 [12]. In our case, as B -+ 00, the bootstrap variance estimator Ut(B) computed from the aiJ's converge to Ut(oo) := maxj Varp[aij] where aij denotes the res amp led deviation and F denotes the distribution generating it. Furthermore, if F -+ F, Ut (00) converges to the true variance Ui = maxj VarF[aij ] as T -+ 00. This is the case, for example, if the original signal is i.i.d. in time. When the data has time structure, F does not necessarily converge to the generating distribution F of the original signal anymore. Although we cannot neglect this difference completely, it is small enough to use our scheme for the purposes considered in this paper, e.g. in TDSEP, where the aij depend on the variation of the time-lagged covariances Cij(T) of the signals, we can show that their estimators Ctj (T) are unbiased: Furthermore, we can bound the difference t:.ijkl(T,V) = COV p [Ctj ( T), Ckl (v)] COVF [Cij(T),Ckl(V)] between the covariance of the real matrices and their boot- strap estimators as if :3a < 1, M ;::: 1, Vi: ICii (T) I :S M aJLICii(O) I. In our experiments, however, the bias is usually found to be much smaller than this upper bound. 3 3.1 Experiments Comparing the separation error with the uncertainty estimate To show the practical applicability of the resampling idea to ICA, the separation error Ei was compared with the uncertainty Ui . The separation was performed on different artificial 2D mixtures of speech and music signals and different iid data sets of the same variance. To achieve different separation qualities, white gaussian noise of different intensity has been added to the mixtures. 0.7 , - - - - - - - - - - - - - - - - - - - - - - - _ 0.6 ur Uj ~ 0.5 =0.015 ? 0.4 ~ . ~0.3 c ~ 0.2 U. = 0.177 ' - - -' j 0.1 o L---~~~~~~~~--~ o 0.2 0.4 0.6 separation error E j o L-----~----~----~--~ 0.05 0.8 0.15 0.25 0.35 0.45 Figure 1: (a) The probability distribution for the separation error for a small uncertainty is close to zero, for higher uncertainty it spreads over a larger range. (b) The expected error increases with the uncertainty. Figure 1 relates the uncertainty to the separation error for JADE (TDSEP results look qualitatively the same) . In Fig.1 (left) we see the separation error distribution which has a strong peak for small values of our uncertainty measure, whereas for large uncertainties it tends to become flat, i.e. - as also seen from Fig.1 (right) the uncertainty reflects very well the true separation error. 3.2 Selecting the appropriate BSS algorithm As our variance estimation gives a high correlation to the (true) separation error, the next logical step is to use it as a model selection criterion for: (a) selecting some hyperparameter of the BSS algorithm, e.g. choosing the lag values for TDSEP or (b) choosing between a set of different algorithms that rely on different assumptions about the data, i.e. higher order statistics (e.g. JADE, INFO MAX, FastICA, ... ) or second order statistics (e.g. TDSEP). It could, in principle, be much better to extract the first component with one and the next with another assumption/ algorithm. To illustrate the usefulness of our reliability measure, we study a five-channel mixture of two channels of pure white gaussian noise, two audio signals and one channel of uniformly distributed noise. The reliability analysis for higher order statistics (JADE) 0.3 0.25 0.25 ~- 0.2 E 0.15 :rg 0.1 ::J g ::J TDSEP 3 9.17.10- 5 ~- 0.2 E :rg temporal decorrelation (TDSEP) 0.3 0.15 g 0.05 TDSEP 4 1.29.10-5 ,---- ,---- ,---0 .1 0.05 3 ICA Channel i 3 ICA Channel i Figure 2: Uncertainty of leA projections of an artificial mixture using JADE and TDSEP. Resampling displays the strengths and weaknesses of the different models JADE gives the advice to rely only on channels 3,4,5 (d. Fig.2 left). In fact , these are the channels that contain the audio signals and the uniformly distributed noise. The same analysis applied to the TDSEP-projections (time lag = 0, ... ,20) shows, that TDSEP can give reliable estimates only for the two audio sources (which is to be expected; d. Fig.2 right). According to our measure, the estimation for the audio sources is more reliable in the TDSEP-case. Calculation of the separation error verifies this: TDSEP separates better by about 3 orders of magnitude (JADE: E3 = 1.5 . 10- 1 , E4 = 1.4 . 10- 1 , TDSEP: E 3 = 1.2 . 10- 4 , E4 = 8.7? 10- 5 ). Finally, in our example, estimating the audio sources with TDSEP and after this applying JADE to the orthogonal subspace, gives the optimal solution since it combines the small separation errors E 3, E4 for TDSEP with the ability of JADE to separate the uniformly distributed noise. 3.3 Blockwise uncertainty estimates For a longer time series it is not only important to know which ICA channels are reliable, but also to know whether different parts of a given time series are more (or less) reliable to separate than others. To demonstrate these effects, we mixed two audio sources (8kHz, lOs - 80000 data points) , where the mixtures are partly corrupted by white gaussian noise. Reliability analysis is performed on windows of length 1000, shifted in steps of 250; the resulting variance estimates are smoothed. Fig.3 shows again that the uncertainty measure is nicely correlated with the true separation error, furthermore the variance goes systematically up within the noisy part but also in other parts of the time series that do not seem to match the assumptions underlying the algorithm. 3 So our reliability estimates can eventually Figure 3: Upper panel: mixtures, partly corrupted by noise. Lower panel: the blockwise variance estimate (solid line) vs the true separation error on this block (dotted line) . be used to improve separation performance by removing all but the 'reliable' parts of the time series. For our example this reduces the overall separation error by 2 orders of magnitude from 2.4.10- 2 to 1.7.10- 4 . This moving-window resampling can detect instabilities of the projections in two different ways: Besides the resampling variance that can be calculated for each window, one can also calculate the change of the projection directions between two windows. The later has already been used successfully by Makeig et. al. [10]. 4 Assigning Meaning: Application to Biomedical Data We now apply our reliability analysis to biomedical data that has been produced by an MEG experiment with acoustic stimulation. The stimulation was achieved by presenting alternating periods of music and silence, each of 30s length, to the subjects right ear during 30 min. of total recording time (for details see [13]). The measured DC magnetic field values, sampled at a frequency of 0.4 Hz , gave a total number of 720 sample points for each of the 49 channels. While previously 3For example, the peak in the last third of the time series can be traced back to the fact that the original time series are correlated in this region. [13] analysing the data, we found that many of the ICA components are seemingly meaningless and it took some medical knowledge to find potential meaningful projections for a later close inspection. However, our reliability assessment can also be seen as indication for meaningful projections, i.e. meaningful components should have low variance. In the experiment, BSS was performed on the 23 most powerful principal components using (a) higher order statistics (JADE) and (b) temporal decorrelation (TDSEP, time lag 0 .. 50). The results in Fig.4 show that none of higher order statistics (JADE) temporal decorrelation (TDSEP) 0.35 0.35 0.3 0.3 0.25 0 .25 - ::J ::J ~ 0.2 ~ 0.2 ig ~ g0. 15 0.1 5 ::J ::J 0.1 0 .1 0.05 0.05 10 15 leA-Channel i ,~ 20 10 15 leA-Channel i 20 Figure 4: Resampling on the biomedical data from MEG experiment shows: (a) no JADE projection is reliable (has low uncertainty) (b) TDSEP is able to identify three sources with low uncertainty. the JADE-projections (left) have small variance whereas TDSEP (right) identifies three sources with a good reliability. In fact , these three components have physical meaning: while component 23 is an internal very low frequency signal (drift) that is always present in DC-measurements, component 22 turns out to be an artifact of the measurement; interestingly component 6 shows a (noisy) rectangular waveform that clearly displays the 1/308 on/off characteristics of the stimulus (correlation to stimulus 0.7; see Fig.5) . The clear dipole-structure of the spatial field pattern in 0.5 ~ ~O In ~ stimulUS -0.5 1 234 5 6 7 t[min) Figure 5: Spatial field pattern, frequency content and time course of TDSEP channel 6. Fig.5 underlines the relevance of this projection. The components found by JADE do not show such a clear structure and the strongest correlation of any component to the stimulus is about 0.3, which is of the same order of magnitude as the strongest correlated PCA-component before applying JADE. 5 Discussion We proposed a simple method to estimate the reliability of ICA projections based on res amp ling techniques. After showing that our technique approximates the separation error, several directions are open(ed) for applications. First, we may like to use it for model selection purposes to distinguish between algorithms or to chose appropriate hyperparameter values (possibly even component-wise). Second, variances can be estimated on blocks of data and separation performance can be enhanced by using only low variance blocks where the model matches the data nicely. Finally reliability estimates can be used to find meaningful components. Here our assumption is that the more meaningful a component is, the more stably we should be able to estimate it. In this sense artifacts appear of course also as meaningful, whereas noisy directions are discarded easily, due to their high uncertainty. Future research will focus on applying res amp ling techniques to other unsupervised learning scenarios. We will also consider Bayesian modelings where often a variance estimate comes for free, along with the trained model. Acknowledgments K-R.M thanks Guido Nolte and the members of the Oberwolfach Seminar September 2000 in particular Lutz Dumbgen and Enno Mammen for helpful discussions and suggestions. K -R. M and A. Z. acknowledge partial funding by the EU project (IST-1999-14190 - BLISS). We thank the Biomagnetism Group of the PhysikalischTechnische Bundesanstalt (PTB) for providing the MEG-DC data. References [1] S. Amari, A. Cichocki, and H. H. Yang. A new learning algorithm for blind signal separation. In D .S. Touretzky, M.C. Mozer, and M.E . Hasselmo , editors, Advances in Neural Information Processing Systems (NIPS 95), volume 8, pages 882-893. The MIT Press, 1996. [2] A. J. Bell and T. J. Sejnowski. An information maximisation approach to blind separation and blind deconvolution. N eural Computation, 7:1129- 1159, 1995. [3] R. Beran and M.S. Srivastava. Bootstrap tests and confidence regions for functions of a covariance matrix. Annals of Statistics, 13:95- 115, 1985. [4] J.-F. Cardoso and A. Souloumiac. Blind beamforming for non Gaussian signals. IEEE Proceedings-F, 140(6):362- 370, December 1994. [5] P. Comon. Independent component analysis, a new concept ? 36(3):287-314, 1994. Signal Processing, [6] G. Deco and D. Obradovic. An information-theoretic approach to neural computing. Springer, New York, 1996. [7] B. Efron and R.J. Tibshirani. An Introduction to the Bootstrap. Chapman & Hall, first edition, 1993. [8] A. Hyviirinen, J. Karhunen , and E. Oja. Independent Compon ent Analysis. Wiley, 200l. [9] Ch. Jutten and J. Herault. Blind separation of sources, part I: An adaptive algorithm based on neuromimetic architecture . Signal Processing, 24:1- 10, 1991. [10] S. Makeig, S. Enghoff, T.-P. Jung, and T. Sejnowski. Moving-window ICA decomposition of EEG data reveals event-related changes in oscillatory brain activity. In Proc. 2nd Int. Workshop on Independent Component Analysis and Blind Source Separation (ICA '2000), pages 627- 632 , Helsinki, Finland, 2000. [11] F . Meinecke, A. Ziehe, M. Kawanabe, and K-R. Muller. Assessing reliability of ica projections - a resampling approach. In ICA '01. T.-W. Lee, Ed., 200l. [12] J. Shao and D. Th. The Jackknife and Bootstrap. Springer, New York, 1995. [13] G. Wubbeler, A. Ziehe, B.-M. Mackert, K-R. Muller, L. Trahms, and G. Curio. Independent component analysis of non-invasively recorded cortical magnetic dc-fields in humans. IEEE Transactions on Biomedical Engineering, 47(5):594-599 , 2000. [14] A. Ziehe and K-R. Muller. TDSEP - an efficient algorithm for blind separation using time structure. In L. Niklasson , M. Boden, and T. Ziemke, editors , Proc. Int. Conf. on Artificial Neural N etworks (ICANN'9S) , pages 675 - 680 , Skiivde, Sweden, 1998. Springer Verlag.
2086 |@word underline:2 nd:1 open:1 tried:1 covariance:3 decomposition:1 solid:1 series:7 selecting:3 r5t:1 interestingly:1 amp:6 ida:1 comparing:2 assigning:1 ij1:1 wx:2 resampling:21 v:1 parameterization:1 plane:1 inspection:1 five:1 along:2 become:1 supply:1 ziemke:1 combine:1 expected:2 ica:17 brain:1 ptb:1 inspired:1 actual:1 str:1 window:5 project:1 estimating:2 underlying:2 bounded:1 panel:2 transformation:2 temporal:5 remember:1 stationarities:1 tackle:1 makeig:2 medical:1 unreliability:1 producing:1 appear:1 before:1 understood:1 engineering:1 tends:1 limit:1 might:2 chose:1 studied:1 range:1 statistically:1 practical:1 acknowledgment:1 yj:1 maximisation:1 block:6 bootstrap:16 procedure:1 bell:1 significantly:1 projection:22 word:1 confidence:1 get:1 cannot:1 close:2 selection:2 applying:4 impossible:1 instability:2 go:1 rectangular:1 pure:1 dipole:1 estimator:4 stability:1 coordinate:2 variation:1 analogous:1 annals:1 enhanced:1 guido:1 element:2 observed:2 calculate:5 region:2 eu:1 principled:1 mozer:1 ui:4 trained:1 depend:1 ithe:1 dilemma:1 icii:1 basis:3 completely:1 shao:1 easily:2 joint:1 various:1 separated:2 sejnowski:2 artificial:3 klaus:1 jade:16 choosing:2 lag:3 larger:1 say:1 amari:1 ability:1 statistic:10 cov:1 noisy:3 seemingly:1 indication:1 took:1 propose:1 ckl:2 mixing:4 achieve:1 description:1 los:1 ent:1 tdsep:24 assessing:2 produce:3 generating:2 converges:1 illustrate:2 oo:1 measured:1 ij:1 strong:1 come:1 direction:9 waveform:1 correct:1 pea:1 vx:1 human:1 enable:1 generalization:1 mab:1 around:1 considered:1 hall:1 algorithmic:1 finland:1 enno:1 purpose:2 estimation:5 proc:2 superposition:1 hasselmo:1 eisi:1 successfully:1 reflects:1 mit:1 clearly:1 always:2 gaussian:4 ctj:2 focus:1 obradovic:1 am:1 detect:1 sense:1 helpful:1 dependent:1 typically:1 fhg:1 germany:2 overall:1 herault:1 arccos:1 spatial:2 special:1 neuen:1 field:4 nicely:2 chapman:1 r5:2 look:1 unsupervised:5 future:1 others:2 stimulus:4 randomly:1 oja:1 maxj:3 attempt:1 interest:1 weakness:1 mixture:8 analyzed:1 partial:1 respective:2 sweden:1 orthogonal:3 re:6 theoretical:1 delete:1 column:2 modeling:3 applicability:1 deviation:1 usefulness:2 fastica:1 inadequate:1 reported:1 answer:1 corrupted:2 thanks:1 peak:2 stay:1 meinecke:3 off:1 lee:1 enhance:1 squared:1 again:2 ear:1 deco:1 recorded:1 choose:1 possibly:1 conf:1 account:1 potential:2 de:1 accompanied:1 coefficient:1 bliss:1 int:2 blind:9 ad:1 vi:1 performed:3 view:1 later:2 recover:2 parallel:1 encephalography:1 ass:3 variance:17 characteristic:1 identify:1 bayesian:1 ijkl:1 produced:1 iid:2 none:1 straight:1 oscillatory:1 strongest:2 touretzky:1 ed:3 frequency:3 sampled:1 popular:1 logical:1 knowledge:2 ut:3 efron:1 back:1 higher:6 varf:1 strongly:1 furthermore:3 just:1 biomedical:5 varp:1 correlation:4 ei:4 assessment:1 lack:1 defines:1 jutten:1 stably:1 artifact:2 reveal:1 quality:2 effect:1 contain:1 true:7 unbiased:1 concept:1 vicinity:1 alternating:1 white:3 attractive:1 during:1 mammen:1 whiten:1 oberwolfach:1 criterion:1 palais:1 presenting:1 theoretic:1 demonstrate:2 passive:1 fj:3 meaning:3 wise:3 consideration:1 funding:1 rotation:12 niklasson:1 stimulation:2 physical:2 khz:1 volume:1 approximates:1 measurement:2 reliability:15 moving:2 specification:1 longer:1 whitening:1 scenario:1 verlag:1 muller:3 seen:2 converge:2 period:1 signal:17 relates:1 full:2 desirable:1 enghoff:1 reduces:1 match:2 calculation:1 cross:3 basic:1 whitened:1 achieved:1 lea:7 whereas:3 addressed:1 source:21 meaningless:1 subject:1 recording:1 hz:1 member:1 beamforming:1 december:1 seem:1 yang:1 enough:1 xj:1 affect:1 gave:1 nolte:1 competing:1 nabe:1 architecture:1 idea:1 whether:2 pca:3 miiller:1 speech:1 e3:1 york:2 remark:1 repeatedly:1 clear:2 cardoso:1 generate:1 canonical:1 shifted:1 dotted:1 estimated:4 tibshirani:1 hyperparameter:2 group:4 key:1 ist:1 traced:1 drawn:1 angle:5 parameterized:1 fourth:1 telecommunication:1 uncertainty:17 powerful:1 separation:34 draw:1 scaling:1 bound:2 datum:2 display:2 correspondence:1 fold:2 distinguish:1 yielded:1 activity:1 strength:1 occur:2 helsinki:1 flat:1 min:2 performing:1 jackknife:4 according:1 remain:1 smaller:1 ur:1 modification:1 comon:1 invariant:1 covf:1 previously:1 etworks:1 turn:1 eventually:2 needed:2 know:2 neuromimetic:1 apply:2 kawanabe:2 appropriate:3 magnetic:2 anymore:1 original:5 denotes:2 instant:1 neglect:1 music:2 uj:1 tensor:1 objective:1 g0:1 question:2 already:3 added:1 surrogate:8 september:1 subspace:1 separate:4 thank:1 berlin:1 separating:3 whom:1 enforcing:1 meg:4 besides:2 length:3 providing:1 cij:3 abelian:1 potentially:1 blockwise:2 info:1 lagged:2 unknown:4 perform:1 upper:2 boot:1 discarded:1 acknowledge:1 dc:4 discovered:1 smoothed:1 intensity:1 drift:1 acoustic:2 potsdam:2 nip:1 able:2 bar:3 usually:1 pattern:2 summarize:1 reliable:7 max:1 deleting:1 event:1 decorrelation:5 rely:2 scheme:1 improve:2 brief:1 identifies:1 fraunhofer:1 extract:1 cichocki:1 lutz:1 literature:1 asymptotic:1 mixed:1 generation:1 suggestion:1 var:2 generator:1 validation:3 consistent:1 principle:1 editor:2 systematically:1 beran:1 course:4 jung:1 last:1 free:1 aij:13 bias:1 allow:1 silence:1 telling:1 wide:1 distributed:4 slice:1 bs:13 calculated:1 souloumiac:1 cortical:1 forward:1 made:1 commonly:1 qualitatively:1 ig:1 adaptive:1 far:1 transaction:1 sj:1 approximate:1 reveals:1 conclude:1 xi:3 don:1 hyviirinen:1 channel:16 correlated:4 obtaining:1 eeg:1 boden:1 excellent:1 necessarily:1 domain:1 icann:1 spread:2 r5b:1 noise:8 ling:5 edition:1 verifies:1 repeated:1 body:1 eural:1 fig:8 advice:1 wiley:1 seminar:1 third:1 e4:3 removing:1 showing:1 invasively:1 demixing:1 deconvolution:1 workshop:1 curio:1 diagonalization:1 magnitude:3 karhunen:1 rg:2 led:1 simply:1 expressed:1 springer:3 mij:1 ch:1 relies:1 goal:2 identity:1 content:1 experimentally:1 change:3 analysing:1 uniformly:3 miss:1 principal:1 strap:1 called:1 total:2 partly:2 experimental:1 meaningful:7 select:1 ziehe:5 internal:1 mark:1 cumulant:1 relevance:1 audio:6 srivastava:1
1,192
2,087
Face Recognition Using Kernel Methods Ming-Hsuan Yang Honda Fundamental Research Labs Mountain View, CA 94041 myang@hra.com Abstract Principal Component Analysis and Fisher Linear Discriminant methods have demonstrated their success in face detection, recognition, and tracking. The representation in these subspace methods is based on second order statistics of the image set, and does not address higher order statistical dependencies such as the relationships among three or more pixels. Recently Higher Order Statistics and Independent Component Analysis (ICA) have been used as informative low dimensional representations for visual recognition. In this paper, we investigate the use of Kernel Principal Component Analysis and Kernel Fisher Linear Discriminant for learning low dimensional representations for face recognition, which we call Kernel Eigenface and Kernel Fisherface methods. While Eigenface and Fisherface methods aim to find projection directions based on the second order correlation of samples, Kernel Eigenface and Kernel Fisherface methods provide generalizations which take higher order correlations into account. We compare the performance of kernel methods with Eigenface, Fisherface and ICA-based methods for face recognition with variation in pose, scale, lighting and expression. Experimental results show that kernel methods provide better representations and achieve lower error rates for face recognition. 1 Motivation and Approach Subspace methods have been applied successfully in numerous visual recognition tasks such as face localization, face recognition, 3D object recognition, and tracking. In particular, Principal Component Analysis (PCA) [20] [13] ,and Fisher Linear Discriminant (FLD) methods [6] have been applied to face recognition with impressive results. While PCA aims to extract a subspace in which the variance is maximized (or the reconstruction error is minimized), some unwanted variations (due to lighting, facial expressions, viewing points, etc.) may be retained (See [8] for examples). It has been observed that in face recognition the variations between the images of the same face due to illumination and viewing direction are almost always larger than image variations due to the changes in face identity [1]. Therefore, while the PCA projections are optimal in a correlation sense (or for reconstruction" from a low dimensional subspace), these eigenvectors or bases may be suboptimal from the classification viewpoint. Representations of Eigenface [20] (based on PCA) and Fisherface [6] (based on FLD) methods encode the pattern information based on the second order dependencies, i.e., pixelwise covariance among the pixels, and are insensitive to the dependencies among multiple (more than two) pixels in the samples. Higher order dependencies in an image include nonlinear relations among the pixel intensity values, such as the relationships among three or more pixels in an edge or a curve, which can capture important information for recognition. Several researchers have conjectured that higher order statistics may be crucial to better represent complex patterns. Recently, Higher Order Statistics (HOS) have been applied to visual learning problems. Rajagopalan et ale use HOS of the images of a target object to get a better approximation of an unknown distribution. Experiments on face detection [16] and vehicle detection [15] show comparable, if no better, results than other PCA-based methods. The concept of Independent Component Analysis (ICA) maximizes the degree of statistical independence of output variables using contrast functions such as Kullback-Leibler divergence, negentropy, and cumulants [9] [10]. A neural network algorithm to carry out ICA was proposed by Bell and Sejnowski [7], and was applied to face recognition [3]. Although the idea of computing higher order moments in the ICA-based face recognition method is attractive, the assumption that the face images comprise of a set of independent basis images (or factorial codes) is not intuitively clear. In [3] Bartlett et ale showed that ICA representation outperform PCA representation in face recognition using a subset of frontal FERET face images. However, Moghaddam recently showed that ICA representation does not provide significant advantage over PCA [12]. The experimental results suggest that seeking non-Gaussian and independent components may not necessarily yield better representation for face recognition. In [18], Sch6lkopf et ale extended the conventional PCA to Kernel Principal Component Analysis (KPCA). Empirical results on digit recognition using MNIST data set and object recognition using a database of rendered chair images showed that Kernel PCA is able to extract nonlinear features and thus provided better recognition results. Recently Baudat and Anouar, Roth and Steinhage, and Mika et ale applied kernel tricks to FLD and proposed Kernel Fisher Linear Discriminant (KFLD) method [11] [17] [5]. Their experiments showed that KFLD is able to extract the most discriminant features in the feature space, which is equivalent to extracting the most discriminant nonlinear features in the original input space. In this paper we seek a method that not only extracts higher order statistics of samples as features, but also maximizes the class separation when we project these features to a lower dimensional space for efficient recognition. Since much of the important information may be contained in the high order dependences among the pixels of a: face image, we investigate the use of Kernel PCA and Kernel FLD for face recognition, which we call Kernel Eigenface and Kernel Fisherface methods, and compare their performance against the standard Eigenface, Fisherface and ICA methods. In the meanwhile, we explain why kernel methods are suitable for visual recognition tasks such as face recognition. 2 Kernel Principal Component Analysis Given a set of m centered (zero mean, unit variance) samples Xk, Xk == [Xkl, ... ,Xkn]T ERn, PCA aims to find the projection directions that maximize the variance, C, which is equivalent to finding the eigenvalues from the covariance matrix AW=CW (1) for eigenvalues A ~ 0 and eigenvectors W E Rn. In Kernel PCA, each vector x is projected from the input space, Rn, to a high dimensional feature space, Rf, by a nonlinear mapping function: <t> : Rn -+ Rf, f ~ n. Note that the dimensionality of the feature space can be arbitrarily large. In Rf, the corresponding eigenvalue problem is "AW4> = C4>w4> (2) where C4> is a covariance matrix. All solutions weI> with A =I- 0 lie in the span of <t> (x1), ..., <t> (Xm ), and there exist coefficients ai such that m w4> = E ai<t>(xi) (3) i=l Denoting an m x m matrix K by K?? x?) -- <t>(x?)? <t>(x?) ~1 - k(x?~'1 ~ 1 (4) , the Kernel PCA problem becomes mAKa =K2 a (5) (6) mAa =Ka where a denotes a column vector with entries aI, ... , am. The above derivations assume that all the projected samples <t>(x) are centered in Rf. See [18] for a ~ethod to center the vectors <t>(x) in Rf. Note that conventional PCA is a special case of Kernel PCA with polynomial kernel of first order. In other words, Kernel PCA is a generalization of conventional PCA since different kernels can be utilized for different nonlinear projections. We can now project the vectors in Rf to a lower dimensional space spanned by the eigenvectors weI>, Let x be a test sample whose projection is <t>(x) in Rf, then the projection of <t>(x) onto the eigenvectors weI> is the nonlinear principal components corresponding to <t>: m w4> . <t>(x) = E ai (<t> (Xi) . <t>(x)) m = i=l E aik(xi, x) (7) i=l In other words, we can extract the first q (1 ~ q ~ m) nonlinear principal components (Le., eigenvectors w4? using the kernel function without the expensive operation that explicitly projects the samples to a high dimensional space Rf" The first q components correspond to the first q non-increasing eigenvalues of (6). For face recognition where each x encodes a face image, we call the extracted nonlinear principal components Kernel Eigenfaces. 3 Kernel Fisher Linear Discriminant Similar to the derivations in Kernel PCA, we assume the projected samples <t>(x) are centered in Rf (See [18] for a method to center the vectors <t>(x) in Rf), we formulate the equations in a way that use dot products for FLD only. Denoting the within-class and between-class scatter matrices by S~ and SiJ, and applying FLD in kernel space, we need to find eigenvalues A and eigenvectors weI> of AS~WeI> = siJweI> (8) , which can be obtained by <P WOPT I(W<P)T S~W<P I = argw;x I(Wq,)TS~Wq,1 = [<P Wl <P W2 ... w;.] (9) where {w[ Ii == 1, 2, ... ,m} is the set of generalized eigenvectors corresponding to the m largest generalized eigenvalues {Ai Ii == 1,2, ... ,m}. For given classes t and u and their samples, we define the kernel function by Let K be a m x m matrix defined by the elements (Ktu)~1;:::,cc, where K tu is a matrix composed of dot products in the feature space Rf, Le., K == (Ktu )=l, u=l, ,c,c where K tu == (k rs )r=l, s=l, ,lt ,I'U (11) Note K tu is a It x Iu matrix, and K is a m x m symmetric matrix. We also define a matrix Z: (12) where (Zt) is a It x It matrix with terms all equal to ~, Le., Z is a m x m block diagonal matrix. The between-class and within-class scatter matrices in a high dimensional feature space Rf are defined as c siJ == L liJ.ti (p/f)T (13) ep(Xij )~(Xij)T (14) i=l C Ii i=l j=l LL S~ == where pi is the mean of class i in Rf, Ii is the number of samples belonging to class i. From the theory of reproducing kernels, any solution w<P E Rf must lie in the span of all training samples in Rf, Le., c w<P == Ip LL cy'pqep(xpq ) (15) p=lq=l It follows that we can get the solution for (15) by solving: AKKa==KZKa (16) Consequently, we can write (9) as <P I(WifJ)T sifJwifJl WOPT == argmaxwifJ I(WifJ)TS!WifJ I == argmaxw?p == [wi ... w~] laKZKal laKKal (17) We can project ~(x) to a lower dimensional space spanned by the eigenvectors w<P in a way similar to Kernel PCA (See Section 2). Adopting the same technique in the Fisherface method (which avoids singularity problems in computing W6PT) for face recognition [6], we call the extracted eigenvectors in (17) Kernel Fisherfaces. 4 Experiments We test both kernel methods against standard rCA, Eigenface, and Fisherface methods using the publicly available AT&T and Yale databases. The face images in these databases have several unique characteristics. While the images in the AT&T database contain the facial contours and vary in pose as well scale, the face images in the Yale database have been cropped and aligned. The face images in the AT&T database were taken under well controlled lighting conditions whereas the images in the Yale database were acquired under varying lighting conditions. We use the first database as a baseline study and then use the second one to evaluate face recognition methods under varying lighting conditions. 4.1 Variation in Pose and Scale The AT&T (formerly Olivetti) database contains 400 images of 40 subjects. To .reduce computational complexity, each face image is downsampled to 23 x 28 pixels. We represent each image by a raster scan vector of the intensity values, .and then normalize them to be zero-mean vectors. The mean and standard deviation of Kurtosis of the face images are 2.08 and 0.41, respectively (the Kurtosis of a Gaussian distribution is 3). Figure 1 shows images of two subjects. In contrast to images of the Yale database, the images include the facial contours, and variation in pose as well as scale. However, the lighting conditions remain constant. Fig~re 1: Face images in the AT&T database (Left) and the Yale database (Right). The experiments are performed using the "leave-one-out" strategy: To classify an image of person, that image is removed from the training set of (m - 1) images and the projection matrix is computed. All the m images in the training set are projected to a reduced space using the computed projection matrix w or weI> and recognition is performed based on a nearest neighbor classifier. The number of principal components or independent components are empirically determined to achieve the lowest error rate by each method. Figure 2 shows the experimental results. Among all the methods, the Kernel Fisherface method with Gaussian kernel and second degree polynomial kernel achieve the lowest error rate. Furthermore, the kernel methods perform better than standard rCA, Eigenface and Fisherface methods. Though our experiments using rCA seem to contradict to the good empirical results reported in [3] [4] [2]' a close look at the data sets reveals a significant difference in pose and scale variation of the face images in the AT&T database, whereas a subset of frontal FERET face images with change of expression was used in [3] [2]. Furthermore, the comparative study on classification with respect to PCA in [4] (pp. 819, Table 1) and the errors made by two rCA algorithms in [2] (pp. 50, Figure 2.18) seem to suggest that lCA methods do not have clear advantage over other approaches in recognizing faces with pose and scale variation. 4.2 Variation in Lighting and Expression The Yale database contains 165 images of 11 subjects that includes variation in both facial expression and lighting. For computational efficiency, each image has been downsampled to 29 x 41 pixels. Likewise, each face image is represented by a Method I rCA Eigenface Fisherface Kernel Eigenface, d==2 Kernel Eigenface, d==3 Kernel Fisherface (P) Kernel Fisherface (G) 30 14 50 50 14 14 2.75 (11/400) 1.50 (6/400) 2.50 (10/400) 2.00 (8/400) 1.25 (5/400) 1.25 (5/400) Figure 2: Experimental results on AT&T database. centered vector of normalized intensity values. The mean and standard deviation of Kurtosis of the face images are 2.68 and 1.49, respectively. Figure 1 shows 22 closely cropped images of two subjects which include internal facial structures such as the eyebrow, eyes, nose, mouth and chin, but do not contain the facial contours. Using the same leave-one-out strategy, we experiment with the number of principal components and independent components to achieve the lowest error rates for Eigenface and Kernel Eigenface methods. For Fisherface and Kernel Fisherface methods, we project all the samples onto a subspace spanned by the c - 1 largest eigenvectors. The experimental results are shown in Figure 3. Both kernel methods perform better than standard ICA, Eigenface and Fisherface methods. Notice that the improvement by the kernel methods are rather significant (more than 10%). Notice also that kernel methods consistently perform better than conventional methods for both databases. The performance achieved by the ICA method indicates that face representation using independent sources is not effective when the images are taken under varying lighting conditions. Method 35 30 29.09 28..49 27.27 24.24 ~ 25 I lCA ~ 20 ~ 8 15 ~ 10 o. -< ~ < ul:l-. ~~ ~s -<,-.. ~& Q ....:l ~ Q,-.. ~& ? Q S ~ Eigenface Fisherface Kernel Eigenface, d==2 Kernel Eigenface, d==3 Kernel Fisherface (P) Kernel Fisherface (G) 30 14 80 60 14 14 28.48 (47/165) 8.48 (14/165) 27.27 (45/165) 24.24 (40/165) 6.67 (11/165) 6.06 (10/165) Figure 3: Experimental results on Yale database. Figure 4 shows the training samples of the Yale database projected onto the first two eigenvectors extracted by the Kernel Eigenface and Kernel Fisherface methods. The projected samples of different classes are smeared by the Kernel Eigenface method whereas the samples projected by the Kernel Fisherface are separated quite welL In fact, the samples belonging to the same class are projected to the same position by the largest two eigenvectors. This example provides an explanation to the good results achieved by the Kernel Fisherface method. The experimental results show that Kernel Eigenface and Fisherface methods are able to extract nonlinear features and achieve lower error rate. Instead of using a nearest neighbor classifier, the performance can potentially be improved by other classifiers (e.g., k-nearest neighbor and perceptron). Another potential improvement is to use all the extracted nonlinear components as features (Le., without projecting to a lower dimensional space) and use a linear Support Vector Machine (SVM) to construct a decision surface. Such a two-stage approach is, in spirit, similar to nonlinear SVMs in which the samples are first projected to a high dimensional feature space where a hyperplane with largest hyperplane is constructed. In fact, one important factor of the recent success in SVM applications for visual recognition is due to the use of kernel methods. r1-o+. ? + 1- :.. .?it'CIlIl''''''~ IX'';:) : ; ? ? ~ :.'" ? :.u??.. ? ;.,,? .. ? .. ? ~ -21- ~ , ~ : , , ?,??? .. ???1 ~ ~ ~:::: ~ 1- , ;. : : ; ?, .. ?? y V A 1iiI=-*$*_~" 1- <l :* : : ; : : ; : : ; : : : ; ?.. c1ass4 0 class? class8 class9 I- ; 0 ;: class13 1- , , ." 01- , : -: E ? ~i:~~H 1 ? I- * 1 class 1 ~::::~ : 1- .......... :? ..~ ...... ?~~:-~"'-e>"0~O?"'....???:-?-?.-(I?? .. ? .. ;??????O?.?? .. ; .... ??????-:?? ..........:.. ?? ...... ?1 ~~~~::~::~~~ ~ ~ : 1> ~ : , , :.? * , ? ?;? 1 ? .. 1 ?' .. ?? , * ! 1 '-' <l t> ~* ~ 5 --=.,,,0.:- -1 0 * * '* o 2 4 (a) Kernel Eigenface method. :0.08 -0.06 -0.04 -0.02 0.02 0.04 0.)6 O.DB (b) Kernel Fisherface method. Figure 4: Samples projected by Kernel PCA and Kernel Fisher methods. 5 Discussion and Conclusion The representation in the conventional Eigenface and Fisherface approaches is based on second order statistics of the image set, Le., covariance matrix, and does not use high order statistical dependencies such as the relationships among three or more pixels. For face recognition, much of the important information may be contained in the high order statistical relationships among the pixels. Using the kernel tricks that are often used in SVMs, we extend the conventional methods to kernel space where we can extract nonlinear features among three or more pixels. We have investigated Kernel Eigenface and Kernel Fisherface methods, and demonstrate that they provide a more effective representation for face recognition. Compared to other techniques for nonlinear feature extraction, kernel methods have the advantages that they do not require nonlinear optimization, but only the solution of an eigenvalue problem. Experimental results on two benchmark databases show that Kernel Eigenface and Kernel Fisherface methods achieve lower error rates than the ICA, Eigenface and Fisherface approaches in face recognition. The performance achieved by the ICA method also indicates that face representation using independent basis images is not effective when the images contain pose, scale or lighting variation. Our future work will focus on analyzing face recognition methods using other kernel methods in high dimensional space. We plan to investigate and compare the performance of other face recognition methods [14] [12] [19]. References [1] Y. Adini, Y. Moses, and S. Ullman. Face recognition: The problem of compensating for changes in illumination direction. IEEE PAMI, 19(7):721-732, 1997. [2] M. S. Bartlett. Face Image Analysis by Unsupervised Learning and Redundancy Reduction. PhD thesis, University of California at San Diego, 1998. [3] M. S. Bartlett, H. M. Lades, and T. J. Sejnowski. Independent component representations for face recognition. In Proc. of SPIE, volume 3299, pages 528-539, 1998. [4] M. S. Bartlett and T. J. Sejnowski. Viewpoint invariant face recognition using independent component analysis and attractor networks. In NIPS 9, page 817, 1997. [5] G. Baudat and F. Anouar. Generalized discriminant analysis using a kernel approach. Neural Computation, 12:2385-2404,2000. [6] P. Belhumeur, J. Hespanha, and D. Kriegman. Eigenfaces vs. Fisherfaces: Recognition using class specific linear projection. IEEE PAMI, 19(7):711-720, 1997. [7] A. J. Bell and T. J. Sejnowski. An information - maximization approach to blind separation and blind deconvolution. Neural Computation, 7(6):11291159, 1995. [8] C. 1\1. Bishop. fleural fretworks for .J.Dattern Recognition. Oxford University Press, 1995. [9] P. Comon. Independent component analysis: A new concept? Signal Processing, 36(3):287-314-, 1994. [10] A. Hyviirinen, J. Karhunen, and E. Oja. Independent Component Analysis. Wiley-Interscience, 2001. [11] S. Mika, G. Riitsch, J. Weston, B. Sch6lkopf, A. Smola, and K.-R. Muller. Invariant feature extraction and classification in kernel spaces. In NIPS 12, pages 526-532, 2000. [12] B. Moghaddam. Principal manifolds and bayesian subspaces for visual recognition. In Proc. IEEE Int'l Conf. on Computer Vision, pages 1131-1136,1999. [13] B. Moghaddam and A. Pentland. Probabilistic visual learning for object recognition. IEEE PAMI, 19(7):696-710, 1997. [14] P. J. Phillips. Support vector machines applied to face recognition. In NIPS 11, pages 803-809, 1998. [15] A. N. Rajagopalan, P. Burlina, and R. Chellappa. Higher order statistical learning for vehicle detection in images. In Proc. IEEE Int'l Con!. on Computer Vision, volume 2, pages 1204-1209,1999. [16] A. N. Rajagopalan, K. S. Kumar, J. Karlekar, R. Manivasakan, and M. M. Patil. Finding faces in photographs. In Proc. IEEE Int'l Conf. on Computer Vision, pages 640-645, 1998. [17] V. Roth and V. Steinhage. Nonlinear discriminant analysis using kernel functions. In NIPS 12, pages 568-574,2000. [18] B. Sch6lkopf, A. Smola, and K.-R. Muller. Nonlinear component analysis as a kernel eigenvalue problem. Neural Computation, 10(5):1299-1319,1998. [19] Y. W. Teh and G. E. Hinton. Rate-coded restricted Boltzmann machines for face recognition. In NIPS 13, pages 908-914, 2001. [20] M. Turk and A. Pentland. Eigenfaces for recognition. J. of Cognitive Neuroscience, 3(1):71-86, 1991.
2087 |@word polynomial:2 r:1 seek:1 covariance:4 xkn:1 carry:1 reduction:1 moment:1 contains:2 denoting:2 riitsch:1 ka:1 com:1 negentropy:1 scatter:2 must:1 informative:1 v:1 xk:2 provides:1 honda:1 constructed:1 fld:6 interscience:1 acquired:1 ica:12 compensating:1 ming:1 increasing:1 becomes:1 provided:1 project:5 maximizes:2 lowest:3 mountain:1 finding:2 anouar:2 ti:1 unwanted:1 k2:1 classifier:3 unit:1 analyzing:1 oxford:1 pami:3 mika:2 unique:1 block:1 digit:1 w4:4 empirical:2 bell:2 projection:9 word:2 downsampled:2 suggest:2 get:2 onto:3 close:1 baudat:2 applying:1 conventional:6 equivalent:2 demonstrated:1 roth:2 center:2 formulate:1 hsuan:1 spanned:3 variation:11 target:1 diego:1 aik:1 trick:2 element:1 recognition:43 expensive:1 utilized:1 database:19 observed:1 ep:1 capture:1 cy:1 removed:1 complexity:1 kriegman:1 solving:1 localization:1 efficiency:1 basis:2 represented:1 derivation:2 separated:1 effective:3 chellappa:1 sejnowski:4 whose:1 quite:1 larger:1 statistic:6 ip:1 advantage:3 eigenvalue:8 kurtosis:3 reconstruction:2 product:2 tu:3 aligned:1 achieve:6 normalize:1 r1:1 comparative:1 leave:2 object:4 pose:7 nearest:3 kzka:1 direction:4 closely:1 centered:4 viewing:2 eigenface:26 require:1 generalization:2 singularity:1 mapping:1 vary:1 proc:4 largest:4 wl:1 successfully:1 smeared:1 always:1 gaussian:3 aim:3 rather:1 varying:3 encode:1 focus:1 improvement:2 consistently:1 indicates:2 contrast:2 baseline:1 sense:1 am:1 relation:1 pixel:11 iu:1 among:10 classification:3 plan:1 special:1 equal:1 comprise:1 construct:1 extraction:2 look:1 unsupervised:1 future:1 minimized:1 oja:1 composed:1 divergence:1 attractor:1 detection:4 investigate:3 moghaddam:3 edge:1 hra:1 facial:6 re:1 column:1 classify:1 cumulants:1 maximization:1 kpca:1 deviation:2 subset:2 entry:1 recognizing:1 pixelwise:1 reported:1 dependency:5 aw:1 person:1 fundamental:1 probabilistic:1 thesis:1 conf:2 cognitive:1 ullman:1 account:1 potential:1 includes:1 coefficient:1 int:3 explicitly:1 blind:2 vehicle:2 view:1 performed:2 lab:1 publicly:1 variance:3 characteristic:1 likewise:1 maximized:1 yield:1 correspond:1 bayesian:1 lighting:10 researcher:1 cc:1 explain:1 against:2 raster:1 pp:2 turk:1 spie:1 con:1 dimensionality:1 higher:9 wei:6 improved:1 though:1 furthermore:2 stage:1 smola:2 correlation:3 nonlinear:16 concept:2 xpq:1 contain:3 normalized:1 lades:1 symmetric:1 leibler:1 attractive:1 ll:2 generalized:3 chin:1 xkl:1 demonstrate:1 image:41 recently:4 empirically:1 insensitive:1 volume:2 extend:1 significant:3 ai:5 phillips:1 fisherface:29 dot:2 impressive:1 surface:1 etc:1 base:1 showed:4 recent:1 olivetti:1 conjectured:1 success:2 arbitrarily:1 muller:2 belhumeur:1 maximize:1 ale:4 ii:4 signal:1 multiple:1 myang:1 coded:1 controlled:1 vision:3 kernel:75 represent:2 adopting:1 achieved:3 cropped:2 whereas:3 source:1 crucial:1 w2:1 subject:4 steinhage:2 db:1 spirit:1 seem:2 call:4 extracting:1 yang:1 iii:1 independence:1 suboptimal:1 reduce:1 idea:1 expression:5 pca:21 bartlett:4 ul:1 clear:2 eigenvectors:12 factorial:1 svms:2 reduced:1 outperform:1 exist:1 xij:2 notice:2 moses:1 neuroscience:1 write:1 redundancy:1 karlekar:1 almost:1 separation:2 decision:1 comparable:1 wopt:2 yale:8 encodes:1 chair:1 span:2 kumar:1 rendered:1 ern:1 lca:2 belonging:2 remain:1 wi:1 feret:2 argmaxw:1 comon:1 intuitively:1 projecting:1 sij:2 rca:5 invariant:2 restricted:1 taken:2 equation:1 nose:1 available:1 operation:1 ho:2 original:1 denotes:1 include:3 patil:1 seeking:1 strategy:2 dependence:1 diagonal:1 subspace:6 cw:1 manifold:1 discriminant:9 code:1 retained:1 relationship:4 potentially:1 hespanha:1 ethod:1 zt:1 boltzmann:1 unknown:1 fisherfaces:2 sch6lkopf:3 perform:3 teh:1 benchmark:1 t:2 pentland:2 extended:1 hinton:1 rn:3 reproducing:1 intensity:3 c4:2 california:1 nip:5 address:1 able:3 pattern:2 xm:1 rajagopalan:3 eyebrow:1 rf:15 explanation:1 mouth:1 suitable:1 eye:1 numerous:1 extract:7 lij:1 formerly:1 degree:2 viewpoint:2 pi:1 perceptron:1 neighbor:3 eigenfaces:3 face:50 ktu:2 curve:1 avoids:1 contour:3 made:1 projected:10 san:1 contradict:1 kullback:1 reveals:1 xi:3 why:1 table:1 hyviirinen:1 ca:1 investigated:1 complex:1 necessarily:1 meanwhile:1 motivation:1 x1:1 fig:1 wiley:1 position:1 lq:1 lie:2 ix:1 specific:1 bishop:1 svm:2 deconvolution:1 mnist:1 phd:1 illumination:2 karhunen:1 lt:1 photograph:1 visual:7 contained:2 tracking:2 maa:1 extracted:4 weston:1 identity:1 consequently:1 fisher:6 change:3 determined:1 hyperplane:2 principal:11 experimental:8 wq:2 internal:1 support:2 scan:1 frontal:2 evaluate:1
1,193
2,088
1 Bayesian morphometry of hippocampal cells suggests same-cell somatodendritic repulsion Giorgio A. Ascoli * Alexei Samsonovich Krasnow Institute for Advanced Study at George Mason University Fairfax, VA 22030-4444 ascoli@gmu.edu asamsono@gmu.edu Abstract Visual inspection of neurons suggests that dendritic orientation may be determined both by internal constraints (e.g. membrane tension) and by external vector fields (e.g. neurotrophic gradients). For example, basal dendrites of pyramidal cells appear nicely fan-out. This regular orientation is hard to justify completely with a general tendency to grow straight, given the zigzags observed experimentally. Instead, dendrites could (A) favor a fixed (?external?) direction, or (B) repel from their own soma. To investigate these possibilities quantitatively, reconstructed hippocampal cells were subjected to Bayesian analysis. The statistical model combined linearly factors A and B, as well as the tendency to grow straight. For all morphological classes, B was found to be significantly positive and consistently greater than A. In addition, when dendrites were artificially re-oriented according to this model, the resulting structures closely resembled real morphologies. These results suggest that somatodendritic repulsion may play a role in determining dendritic orientation. Since hippocampal cells are very densely packed and their dendritic trees highly overlap, the repulsion must be cellspecific. We discuss possible mechanisms underlying such specificity. 1 I n t r od uc t i on The study of brain dynamics and development at the cellular level would greatly benefit from a standardized, accurate and yet succinct statistical model characterizing the morphology of major neuronal classes. Such model could also provide a basis for simulation of anatomically realistic virtual neurons [1]. The model should accurately distinguish among different neuronal classes: a morphological difference between classes would be captured by a difference in model parameters and reproduced in generated virtual neurons. In addition, the model should be self-consistent: there should be no statistical difference in model parameters measured from real neurons of a given class and from virtual neurons of the same class. The assumption that a simple statistical model of this sort exists relies on the similarity of average environmental and homeostatic conditions encountered by individual neurons during development and on the limited amount of genetic information that underlies differentiation of neuronal classes. Previous research in computational neuroanatomy has mainly focused on the topology and internal geometry of dendrites (i.e., the properties described in ?dendrograms?) [2,3]. Recently, we attempted to include spatial orientation in the models, thus generating 2 virtual neurons in 3D [4]. Dendritic growth was assumed to deviate from the straight direction both randomly and based on a constant bias in a given direction, or ?tropism?. Different models of tropism (e.g. along a fixed axis, towards a plane, or away from the soma) had dramatic effects on the shape of virtual neurons [5]. Our current strategy is to split the problem of finding a statistical model describing neuronal morphology in two parts. First, we maintain that the topology and the internal geometry of a particular dendritic tree can be described independently of its 3D embedding (i.e., the set of local dendritic orientations). At the same time, one and the same internal geometry (e.g., the experimental dendrograms obtained from real neurons) may have many equally plausible 3D embeddings that are statistically consistent with the anatomical characteristics of that neuronal class. The present work aims at finding a minimal statistical model describing local dendritic orientation in experimentally reconstructed hippocampal principal cells. Hippocampal neurons have a polarized shape: their dendrites tend to grow from the soma as if enclosed in cones. In pyramidal cells, basal and apical dendrites invade opposite hemispaces (fig. 1A), while granule cell dendrites all invade the same hemispace. This behavior could be caused by a tendency to grow towards the layers of incoming fibers to establish synapses. Such tendency would correspond to a tropism in a direction roughly parallel to the cell main axis. Alternatively, dendrites could initially stem in the appropriate (possibly genetically determined) directions, and then continue to grow approximately in a radial direction from the soma. A close inspection of pyramidal (basal) trees suggests that dendrites may indeed be repelled from their soma (Fig. 1B). A typical dendrite may reorient itself (arrow) to grow nearly straight along a radius from the soma. Remarkably, this happens even after many turns, when the initial direction is lost. Such behavior may be hard to explain without tropism. If the deviations from straight growth were random, one should be able to ?remodel?th e trees by measuring and reproducing the statistics of local turn angles, assuming its independence of dendritic orientation and location. Figure 1C shows the cell from 1A after such remodeling. In this case basal and apical dendrites retain only their initial (stemming) orientations from the original data. The resulting ?cotton ball?s uggests that dendritic turns are not in dependent of dendritic orientation. In this paper, we use Bayesian analysis to quantify the dendritic tropism. 2 M e t h od s Digital files of fully reconstructed rat hippocampal pyramidal cells (24 CA3 and 23 CA1 neurons) were kindly provided by Dr. D. Amaral. The overall morphology of these cells, as well as the experimental acquisition methods, were extensively described [6]. In these files, dendrites are represented as (branching) chains of cylindrical sections. Each section is connected to one other section in the path to the soma, and may be connected on the other extremity to two other sections (bifurcation), one other section (continuation point), or no other section (terminal tip). Each section is described in the file by its ending point coordinates, its diameter and its "parent", i.e., the attached section in the path to the soma [5,7]. In CA3 cells, basal dendrites had an average of 687(?216) continuation points and 72(?17) bifurcations per cell, while apical dendrites had 717(?156) continuation points and 80(?21) bifurcations per cell. CA1 cells had 462(?138) continuation points and 52(?12) bifurcations (basal), 860(?188) continuation points and 120(?22) bifurcations (apical). In the present work, basal and apical trees of CA3 and CA1 pyramidal cells were treated as 4 different classes. Digital data of rat dentate granule cells [8] are kindly made available by Dr. B. Claiborne through the internet (http://cascade.utsa.edu/bjclab). Only 36 of the 42 cells in this archive were used: in 6 cases numerical processing was not accomplished due to minor inconsistencies in the data files. The ?rejected? cells were 1208875, 3319201, 411883, 411884A, 411884B, 803887B. Granule dendrites had 3 549(?186) continuation points and 30(?6) bifurcations per cell. Cells in these or similar formats can be rendered, rotated, and zoomed with a java applet available through the internet (www.cns.soton.ac.uk) [7]. Figure 1: A: A pyramidal cell (c53063) from Amaral?s archive. B: A zoom-in from panel A (arrows point to the same basal tree location). Dotted dendrites are behind the plane. C: Same cell (c53063) with its dendritic orientation remodeled assuming zero tropism and same statistics of all turn angles (see Results). In agreement with the available format of morphological data (described above), the process of dendritic growth1 can be represented as a discrete stochastic process consisting of sequential attachment of new sections to each growing dendrite. Here we keep the given internal geometry of the experimental data while remodeling the 3D embedding geometry (dendritic orientation). The task is to make a remodeled geometry statistically consistent with the original structure. The basic assumption is that neuronal development1 is a Markov process governed by local rules [4]. Specifically, we assume that (i) each step in dendritic outgrowth only depends on the preceding step and on current local conditions; and (ii) dendrites do not undergo geometrical or topological modification after their formation (see, however, Discussion). In this Markov approximation, a plausible 3D embedding can be found by sequentially orienting individual sections, starting from the soma and moving toward the terminals. We are implementing this procedure in two-step iterations (1). First, at a given node i with coordinates ri we select a section i+1, disregard its given orientation, and calculate its most likely expected direction n'i+1 based on the model (here section i+1 connects nodes i and i+1, and n stands for a unit vector). For a continuation point, the most likely direction n'i+1 is computed as the direction of the vector sum ni + vi. The first term is the direction of the parent section ni , and reflects the tendency dendrites exhibit to grow relatively straight due to membrane tension, mechanical properties of the cytoskeleton, etc. The second term is a local value of a vector field: vi = v(ri ), which comprises the influence of external local conditions on the direction of growth (as specified below). Finally, we generate a perturbation of the most likely direction n'i+1 to produce a particular plausible instance of a new direction. In summary, the new direction ni+1 is generated as: 1 Although we resort to a developmental metaphor, our goal is to describe accurately the result of development rather than the process of development. 4 n i +1 = Ti n'i +1 , (1) n'i +1 || n i + v i . Here Ti is an operator that deflects n'i+1 into a random direction. If we view each deflection as a yaw of angle ?i, then the corresponding rolling angle (describing rotation around the axis of the parent dendrite) is distributed uniformly between 0 and 2?. The probability distribution function for deflections as a function of ?i is taken in a form that, as we found, well fits experimental data: P(Ti ) ? e ? ?i ? , (2) where ? << 1 is a parameter of the model. At bifurcation points, the same rule (1), (2) is applied for each daughter independently. A more accurate and plausible description of dendritic orientation at bifurcations might require a more complex model. However, our simple choice yields surprisingly good results (see below). The model (1), (2) can be used in the simulation of virtual neuronal morphology. In this case one would first need to generate the internal geometry of the dendrites [1-5]. Most importantly, model (1), (2) can be used to quantitatively assess the significance of the somatocentric (radial) tropism of real dendrites. Assuming that there is a significant preferential directionality of growth in hippocampal dendrites, the two main alternatives are (see Introduction): HA: The dominating tropic factor is independent of the location of the soma. HB: The dominating tropic factor is radial with respect to the soma. The simplest model for the vector field v that discriminates between these alternative hypotheses includes both factors, A and B, linearly: v i = a + bn ir . (3) Here a = (ax, ay, az) is a constant vector representing global directionality of cellindependent environmental factors (chemical gradients, density of neurites, etc.) influencing dendritic orientation. nri is the unit vector in the direction connecting the soma to node i, thus representing a somatocentric tropic factor. In summary, ax, ay, az, b and ? are the parameters of the model. Finding that the absolute value a = |a| is significantly greater than b would support HA. On the contrary, finding that b is greater than a would support HB. Based on a Bayesian approach, we compute the most likely values of a, b and ? by maximization of the likelihood of all experimentally measured orientations (taken at continuation points only) of a given dendritic tree: (a*, b *) = arg {max ? ? P(Ti ) = arg min {a , b} i a , b ,? } , (4) i where ?i is given by (1)-(3) with experimental section orientations substituted for ni, ni+1, asterisk denotes most likely values, and the average is over all continuation points. Given a* and b*, the value of ?* can be found from the average value of ?i computed with a = a* and b = b*. The relation results from differentiation of (4) by ?. The same relation holds for the average value of ? computed based on the probability distribution function (2) with ? = ?*. Therefore, <?i> computed from the neurometric data with a = a* and b = b* is equal to <?> based on (2) with ? = ?*. The model is thus self-consistent: the measured value of ?* in a remodeled neuron is guaranteed to coincide on average with the input parameter ? used for simulation. In addition, our numerical analysis indicates self-consistency of the model with respect to a and b, when their values are within a practically meaningful range. 5 3 R e s ul t s Results of the Bayesian analysis are presented in Table 1. Parameters a and b were optimized for each cell individually, then the absolute value a = |a| was taken for each cell. The mean value and the standard deviation of a in Table 1 were computed based on the set of the individual absolute values, while each individual value of b was taken with its sign (which was positive in all cases but one). The most likely direction of a varied significantly among cells, i.e., no particular fixed direction was generally preferred. Table 1: Results from Bayesian analysis (mean ? standard deviation). ? is the minimized deflection angle, a and b are parameters of the model (1)-(3) computed according to (4). Dataset ? CA3-bas 16.4 ? 2.3 CA3-apic 15.2 ? 1.9 CA1-bas 16.6 ? 1.6 CA1-apic 19.1 ? 2.0 Granule 19.1 ? 2.7 Original data B a 0.49 ? 0.17 0.08 ? 0.05 0.36 ? 0.16 0.12 ? 0.07 0.49 ? 0.26 0.14 ? 0.10 0.30 ? 0.20 0.16 ? 0.15 1.01 ? 0.64 0.17 ? 0.11 Z coordinate set to zero b A ? 12.0 ? 2.4 0.42 ? 0.15 0.06 ? 0.05 12.0 ? 2.9 0.29 ? 0.23 0.10 ? 0.14 14.2 ? 1.9 0.48 ? 0.31 0.16 ? 0.12 17.3 ? 2.4 0.22 ? 0.17 0.11 ? 0.10 11.0 ? 1.9 0.36 ? 0.16 0.07 ? 0.05 The key finding is that a is not significantly different from zero, while b is significantly positive. The slightly higher coefficient of variation obtained for granule cells could be due to a larger experimental error in the z coordinate (orthogonal to the slice). In several granule cells (but in none of the pyramidal cells) the greater noise in z was apparent upon visual inspection of the rendered structures. Therefore, we re-ran the analysis discarding the z coordinate (right columns). Results changed only minimally for pyramidal cells, and the granule cell parameters became more consistent with the pyramidal cells. The measured average values of the model parameters were used for remodeling of experimental neuronal shapes, as described above. In particular, b was set to 0.5, while a was set to zero. We kept the internal geometry and the initial stemming direction of each tree from the experimental data, and simulated dendritic orientation at all nodes separated by more than 2 steps from the soma. A typical result is shown in Figure 2. Generally, the artificially re-oriented dendrites looked better than one could expect for a model as simple as (1) ? (3). This result may be compared with figure 1C, which shows an example of remodeling based on the same model in the absence of tropism (a = b = 0). Although in this case the shape can be improved by reducing ?, the result never gets as close to a real shape as in Fig. 2 C, D, even when random, uncorrelated local distortions ("shuffling") are applied to the generated geometry. Thus, although the tendency to grow straight represents the dominant component of the model (i.e., b<1), somatocentric tropism may exert a dramatic effect on dendritic shape. Surprisingly, even the asymmetry of the dendritic spread (compare front and side views) is preserved after remodeling. However, two details are difficult to reproduce with this model: the uniform distribution of dendrites in space and other subtle medium-distance correlations among dendritic deflections. In order to account for these properties, we may need to consider spatially correlated inhomogeneities of the medium and possible short range dendrodendritic interactions. 6 4 D i s c us s i on The key results of this work is that, according to Bayesian analysis, dendrites of hippocampal principal cells display a significant radial tropism. This means that the spatial orientation of these neuronal trees can be statistically described as if dendrites were repelled from their own soma. This preferential direction is stronger than any tendency to grow along a fixed direction independent of the location of the soma. These results apply to all dendritic classes, but in general pyramidal cell basal trees (and granule cell dendrites) display a bigger somatocentric tropism than apical trees. Figure 2: Dendritic remodeling with somatocentric tropism. A, B: front and side views of cell 10861 from Amaral' s archive. C, D: Same views after remodeling with parameters       !  #"$ # a = 0, b = 0.5, ? = 0.15 (corresponding to <?> = 17 stem were taken in their original orientations; all subsequent experimental orientations were disregarded and regenerated from scratch according to the model. Assuming that dendrites are indeed repelled from their soma during development, what could be a plausible mechanism? Principal cells are very densely packed in the hippocampus, and their dendrites highly overlap. If repulsion were mediated by a diffusible chemical factor, in order for dendrites to be repelled radially from their own soma, each neuron should have its own specific chemical marker (a fairly unlikely possibility). If the same repulsive factor were released by all neurons, each dendrite would be repelled by hundreds of somata, and not just by their own. The resulting tropism would be perpendicular to the principal cell layer, i.e. each dendrite would be pushed approximately in the same direction, independent of the location of its soma. This scenario is in clear contrast with the result of our statistical analysis. Thus, how can a growing dendrite sense the location of its own soma? One possibility involves the spontaneous spiking activity of neurons during development. A cell that spikes becomes unique in its neighborhood for a short period of time. The philopodia of dendritic growth 7 cones could possess voltage-gated receptors to sense transient chemical gradients (e.g., pH) created by the spiking cell. Only dendrites that are depolarized during the transient chemical gradient (i.e., those belonging to the same spiking cell) would be repelled by it. Alternatively, depolarized philopodia could be sensitive to the small voltage difference created by the spike in the extracellular space (a voltage that can be recorded by tetrodes). The main results obtained with the simple model presented in this work are independent of the z coordinate in the morphometric files, i.e. the most error-prone measurement in the experimental reconstruction. However, it is important to note that any observed deviation of dendritic path from a straight line, including that due to measurement errors, causes an increase in the most likely values of parameters a and b. Another possibility is that dendrites do grow almost precisely in straight lines, and the measured values of a and b reflect distortions of dendritic shapes after development. In order to assess the effect of these factors on a and b, we pre-processed the experimental data by adding a gradually increasing noise to all coordinates of dendritic sections. Then we were able to extrapolate the dependence of a*, b* and <?>* on the amplitude of noise in order to estimate the parameter values in the absence of the experimental error (which was conservatively taken to be of 0.5 ?m). For basal trees of CA3 pyramidal cells, this analysis yielded an estimated ?corrected? value of b between 0.14 and 0.25, with a remaining much smaller than b. Interestingly, our analysis based on extrapolation shows that, regardless of the assumed amount of distortion present in the experimental data, given the numbers measured for CA3 basal trees, positive initial <?> implies positive initial b. In other words, not only measurement errors, but also possible biological distortions of the dendritic tree may not be capable of accounting for the observed positivity of the parameter b. Although these factors affect our results quantitatively, they do not change the statistical significance nor the qualitative trends. However, a more rigorous analysis needs to be carried out. Nevertheless, artificially reoriented dendrites according to our simple model appear almost as realistic as the original structures, and we could not achieve the same result with any choice of parameters in models of distortion without a somatocentric tropism. In conclusion, whether the present Bayesian analysis reveals a phenomenon of somatodendritic repulsion remains an (experimentally testable) open question. What is unquestionable is that the presented model is a significant step forward in the formulation of an accurate statistical description of dendritic morphology. A c k n ow l e d g me n t s This work was supported in part by Human Brain Project Grant R01 NS39600, funded jointly by NINDS and NIMH. References [1] Ascoli G.A. (1999) Progress and perspectives in computational neuroanatomy. Anat. Rec. 257(6):195-207. [2] van Pelt J. (1997) Effect of pruning on dendritic tree topology. J. Theor. Biol. 186(1):17-32. [3] Burke R.E., W. Marks, B. Ulfhake (1992) A parsimonious description of motoneurons dendritic morphology using computer simulation. J. Neurosci. 12(6):2403-2416. [4] Ascoli G.A., J. Krichmar (2000) L-Neuron: a modeling tool for the efficient generation and parsimonious description of dendritic morphology. Neurocomputing 32-33:1003-1011. [5] Ascoli G.A., J. Krichmar, S. Nasuto, S. Senft (2001) Generation, description and storage of dendritic morphology data. Phil. Trans. R. Sci. B, In Press. [6] Ishizuka N., W. Cowan, D. Amaral (1995) A quantitative analysis of the dendritic organization of pyramidal cells in the rat hippocampus. J. Comp. Neurol. 362(1):17-45. [7] Cannon R.C., D. Turner, G. Pyapali, H. Wheal (1998) An on-line archive of reconstructed hippocampal neurons. J Neurosci. Meth. 84(1-2):49-54. [8] Rihn L.L., B. Claiborne (1990) Dendritic growth and regression in rat dentate granule cells during late postnatal development. Dev. Brain Res. 54(1):115-124
2088 |@word cylindrical:1 stronger:1 hippocampus:2 open:1 simulation:4 bn:1 accounting:1 dramatic:2 initial:5 genetic:1 interestingly:1 current:2 od:2 yet:1 must:1 stemming:2 numerical:2 subsequent:1 realistic:2 shape:7 inspection:3 plane:2 postnatal:1 short:2 node:4 location:6 along:3 qualitative:1 expected:1 indeed:2 roughly:1 behavior:2 samsonovich:1 growing:2 morphology:9 brain:3 terminal:2 nor:1 metaphor:1 increasing:1 becomes:1 provided:1 project:1 underlying:1 panel:1 medium:2 what:2 ca1:5 finding:5 differentiation:2 quantitative:1 ti:4 growth:6 uk:1 unit:2 grant:1 appear:2 positive:5 giorgio:1 influencing:1 local:8 receptor:1 extremity:1 path:3 approximately:2 might:1 exert:1 minimally:1 suggests:3 limited:1 range:2 statistically:3 perpendicular:1 unique:1 lost:1 procedure:1 significantly:5 cascade:1 java:1 pre:1 radial:4 regular:1 specificity:1 word:1 suggest:1 get:1 close:2 operator:1 storage:1 influence:1 www:1 phil:1 regardless:1 starting:1 independently:2 focused:1 rule:2 importantly:1 dendrograms:2 embedding:3 coordinate:7 variation:1 spontaneous:1 play:1 hypothesis:1 agreement:1 trend:1 rec:1 observed:3 role:1 calculate:1 connected:2 morphological:3 ran:1 discriminates:1 developmental:1 nimh:1 dynamic:1 upon:1 completely:1 basis:1 represented:2 fiber:1 separated:1 describe:1 formation:1 neighborhood:1 apparent:1 larger:1 plausible:5 dominating:2 distortion:5 favor:1 statistic:2 jointly:1 itself:1 inhomogeneity:1 reproduced:1 reconstruction:1 interaction:1 zoomed:1 krichmar:2 achieve:1 description:5 az:2 parent:3 asymmetry:1 produce:1 generating:1 remodeled:3 rotated:1 ac:1 measured:6 minor:1 progress:1 involves:1 implies:1 quantify:1 direction:23 radius:1 closely:1 stochastic:1 human:1 transient:2 virtual:6 implementing:1 require:1 dendritic:36 biological:1 theor:1 hold:1 practically:1 around:1 burke:1 dentate:2 major:1 dendrodendritic:1 released:1 sensitive:1 individually:1 tool:1 reflects:1 aim:1 rather:1 cannon:1 voltage:3 ax:2 consistently:1 likelihood:1 mainly:1 indicates:1 greatly:1 contrast:1 rigorous:1 sense:2 dependent:1 repulsion:5 unlikely:1 initially:1 relation:2 reproduce:1 reorient:1 overall:1 among:3 orientation:20 arg:2 development:8 spatial:2 bifurcation:8 uc:1 fairly:1 field:3 equal:1 never:1 nicely:1 represents:1 nearly:1 yaw:1 minimized:1 regenerated:1 quantitatively:3 oriented:2 randomly:1 densely:2 zoom:1 individual:4 neurocomputing:1 geometry:9 consisting:1 cns:1 connects:1 maintain:1 organization:1 investigate:1 possibility:4 alexei:1 highly:2 behind:1 chain:1 zigzag:1 accurate:3 capable:1 preferential:2 orthogonal:1 tree:15 gmu:2 re:4 minimal:1 instance:1 column:1 modeling:1 dev:1 measuring:1 maximization:1 ca3:7 apical:6 deviation:4 rolling:1 uniform:1 hundred:1 front:2 combined:1 density:1 retain:1 tip:1 connecting:1 reflect:1 recorded:1 possibly:1 positivity:1 dr:2 external:3 resort:1 account:1 utsa:1 includes:1 coefficient:1 caused:1 depends:1 vi:2 view:4 extrapolation:1 applet:1 sort:1 parallel:1 ass:2 ni:5 ir:1 became:1 characteristic:1 correspond:1 yield:1 bayesian:8 accurately:2 none:1 comp:1 straight:9 explain:1 synapsis:1 acquisition:1 soton:1 dataset:1 radially:1 subtle:1 amplitude:1 neurotrophic:1 higher:1 tension:2 improved:1 formulation:1 rejected:1 just:1 correlation:1 marker:1 invade:2 orienting:1 effect:4 chemical:5 spatially:1 during:5 self:3 branching:1 rat:4 hippocampal:9 ay:2 geometrical:1 recently:1 rotation:1 spiking:3 attached:1 significant:3 neurites:1 measurement:3 shuffling:1 consistency:1 had:5 funded:1 moving:1 similarity:1 etc:2 dominant:1 own:6 perspective:1 rihn:1 scenario:1 continue:1 inconsistency:1 accomplished:1 captured:1 motoneuron:1 george:1 greater:4 preceding:1 neuroanatomy:2 period:1 ii:1 stem:2 equally:1 bigger:1 va:1 underlies:1 basic:1 regression:1 iteration:1 cell:45 preserved:1 morphometry:1 remarkably:1 addition:3 pyramidal:12 grow:10 depolarized:2 posse:1 archive:4 file:5 tend:1 undergo:1 cowan:1 contrary:1 diffusible:1 split:1 embeddings:1 hb:2 independence:1 fit:1 affect:1 topology:3 opposite:1 whether:1 ul:1 cause:1 generally:2 clear:1 amount:2 extensively:1 ph:1 processed:1 diameter:1 simplest:1 continuation:9 http:1 generate:2 dotted:1 sign:1 estimated:1 per:3 anatomical:1 discrete:1 basal:11 key:2 soma:20 nevertheless:1 kept:1 cone:2 sum:1 deflection:4 angle:5 almost:2 parsimonious:2 polarized:1 ninds:1 pushed:1 layer:2 internet:2 guaranteed:1 distinguish:1 display:2 fan:1 topological:1 encountered:1 yielded:1 activity:1 hemispace:1 constraint:1 precisely:1 ri:2 min:1 remodeling:7 rendered:2 format:2 relatively:1 extracellular:1 pelt:1 according:5 ball:1 belonging:1 membrane:2 smaller:1 slightly:1 modification:1 happens:1 unquestionable:1 anatomically:1 gradually:1 taken:6 remains:1 discus:1 describing:3 mechanism:2 turn:4 subjected:1 repulsive:1 available:3 nri:1 apply:1 away:1 appropriate:1 alternative:2 original:5 standardized:1 denotes:1 include:1 remaining:1 testable:1 establish:1 granule:9 r01:1 question:1 looked:1 spike:2 strategy:1 dependence:1 exhibit:1 gradient:4 ow:1 distance:1 fairfax:1 simulated:1 sci:1 me:1 cellular:1 toward:1 neurometric:1 assuming:4 difficult:1 tropic:3 repelled:6 daughter:1 ba:2 packed:2 gated:1 neuron:17 ascoli:5 markov:2 perturbation:1 reproducing:1 homeostatic:1 varied:1 mechanical:1 specified:1 repel:1 optimized:1 cotton:1 trans:1 able:2 below:2 genetically:1 max:1 including:1 overlap:2 treated:1 advanced:1 turner:1 representing:2 meth:1 axis:3 attachment:1 created:2 carried:1 mediated:1 deviate:1 determining:1 fully:1 expect:1 generation:2 krasnow:1 enclosed:1 digital:2 asterisk:1 consistent:5 morphometric:1 uncorrelated:1 prone:1 summary:2 changed:1 surprisingly:2 supported:1 bias:1 side:2 institute:1 characterizing:1 absolute:3 benefit:1 distributed:1 slice:1 van:1 ending:1 stand:1 conservatively:1 forward:1 made:1 coincide:1 amaral:4 reconstructed:4 pruning:1 claiborne:2 preferred:1 keep:1 global:1 sequentially:1 incoming:1 reveals:1 assumed:2 alternatively:2 table:3 scratch:1 correlated:1 dendrite:37 complex:1 artificially:3 substituted:1 kindly:2 significance:2 main:3 spread:1 linearly:2 arrow:2 neurosci:2 noise:3 succinct:1 neuronal:9 fig:3 comprises:1 governed:1 late:1 anat:1 resembled:1 discarding:1 specific:1 mason:1 tropism:14 neurol:1 tetrode:1 exists:1 sequential:1 adding:1 disregarded:1 likely:7 visual:2 environmental:2 relies:1 goal:1 towards:2 absence:2 change:1 hard:2 experimentally:4 directionality:2 determined:2 typical:2 specifically:1 uniformly:1 justify:1 reducing:1 corrected:1 principal:4 tendency:7 experimental:13 attempted:1 disregard:1 meaningful:1 select:1 internal:7 support:2 mark:1 phenomenon:1 biol:1 extrapolate:1
1,194
2,089
Convolution Kernels for Natural Language Michael Collins AT&T Labs?Research 180 Park Avenue, New Jersey, NJ 07932 mcollins@research.att.com Nigel Duffy Department of Computer Science University of California at Santa Cruz nigeduff@cse.ucsc.edu Abstract We describe the application of kernel methods to Natural Language Processing (NLP) problems. In many NLP tasks the objects being modeled are strings, trees, graphs or other discrete structures which require some mechanism to convert them into feature vectors. We describe kernels for various natural language structures, allowing rich, high dimensional representations of these structures. We show how a kernel over trees can be applied to parsing using the voted perceptron algorithm, and we give experimental results on the ATIS corpus of parse trees. 1 Introduction Kernel methods have been widely used to extend the applicability of many well-known algorithms, such as the Perceptron [1], Support Vector Machines [6], or Principal Component Analysis [15]. A key property of these algorithms is that the only operation they require is the evaluation of dot products between pairs of examples. One may therefore replace  into a new the dotproduct with a Mercer kernel, implicitly mapping feature vectors in space , and applying the original algorithm in this new feature space. Kernels provide an efficient way to carry out these calculations when  is large or even infinite. This paper describes the application of kernel methods to Natural Language Processing (NLP) problems. In many NLP tasks the input domain cannot be neatly formulated as a sub set of . Instead, the objects being modeled are strings, trees or other discrete structures which require some mechanism to convert them into feature vectors. We describe kernels for various NLP structures, and show that they allow computationally feasible representations in very high dimensional feature spaces, for example a parse tree representation that tracks all subtrees. We show how a tree kernel can be applied to parsing using the perceptron algorithm, giving experimental results on the ATIS corpus of parses. The kernels we describe are instances of ?Convolution Kernels?, which were introduced by Haussler [10] and Watkins [16], and which involve a recursive calculation over the ?parts? of a discrete structure. Although we concentrate on NLP tasks in this paper, the kernels should also be useful in computational biology, which shares similar problems and structures. 1.1 Natural Language Tasks Figure 1 shows some typical structures from NLP tasks. Each structure involves an ?observed? string (a sentence), and some hidden structure (an underlying state sequence or tree). We assume that there is some training set of structures, and that the task is to learn a) Lou Gerstner is chairman of IBM [S [NP Lou Gerstner ] [VP is [NP chairman [PP of [NP IBM ] ] ] ] ] b) Lou Gerstner is chairman of IBM Lou/SP Gerstner/CP is/N chairman/N of/N IBM/SC c) Lou/N Gerstner/N is/V chairman/N of/P IBM/N Figure 1: Three NLP tasks where a function is learned from a string to some hidden structure. In (a), the hidden structure is a parse tree. In (b), the hidden structure is an underlying sequence of states representing named entity boundaries (SP = Start person, CP = Continue person, SC = Start company, N= No entity). In (c), the hidden states represent part-of-speech tags (N = noun, V = verb, P = preposition,). the mapping from an input string to its hidden structure. We refer to tasks that involve trees as parsing problems, and tasks that involve hidden state sequences as tagging problems. In many of these problems ambiguity is the key issue: although only one analysis is plausible, there may be very many possible analyses. A common way to deal with ambiguity is to use a stochastic grammar, for example a Probabilistic Context Free Grammar (PCFG) for parsing, or a Hidden Markov Model (HMM) for tagging. Probabilities are attached to rules in the grammar ? context-free rules in the case of PCFGs, state transition probabilities and state emission probabilities for HMMs. Rule probabilities are typically estimated using maximum likelihood estimation, which gives simple relative frequency estimates. Competing analyses for the same sentence are ranked using these probabilities. See [3] for an introduction to these methods. This paper proposes an alternative to generative models such as PCFGs and HMMs. Instead of identifying parameters with rules of the grammar, we show how kernels can be used to form representations that are sensitive to larger sub-structures of trees or state sequences. The parameter estimation methods we describe are discriminative, optimizing a criterion that is directly related to error rate. While we use the parsing problem as a running example in this paper, kernels over NLP structures could be used in many ways: for example, in PCA over discrete structures, or in classification and regression problems. Structured objects such as parse trees are so prevalent in NLP that convolution kernels should have many applications. 2 A Tree Kernel The previous section introduced PCFGs as a parsing method. This approach essentially counts the relative number of occurences of a given rule in the training data and uses these counts to represent its learned knowledge. PCFGs make some fairly strong independence assumptions, disregarding substantial amounts of structural information. In particular, it does not appear reasonable to assume that the rules applied at level  in the parse tree are unrelated to those applied at level  . As an alternative we attempt to capture considerably more structural information by considering all tree fragments that occur in a parse tree. This allows us to capture higher order dependencies between grammar rules. See figure 2 for an example. As in a PCFG the new representation tracks the counts of single rules, but it is also sensitive to larger sub-trees. Conceptually we begin by enumerating all tree fragments that occur in the training data       . Note that this is done only implicitly. Each tree is represented by an  dimensional vector where the  ?th component counts the number of occurences of the  ?th tree fragment. Let us define the function   to be the number of occurences of the  ?th tree fragment in tree  , so that  is now represented as     !    "  #$ . a) S NP b) VP N V Jeff ate NP D N the apple NP NP D N the apple D N D N NP the apple D NP N D the N apple Figure 2: a) An example tree. b) The sub-trees of the NP covering the apple. The tree in (a) contains all of these sub-trees, and many others. We define a sub-tree to be any subgraph which includes more than one node, with the restriction that entire (not partial) rule productions must be included. For example, the fragment [NP [D the ]] is excluded because it contains only part of the production NP D N. Note that  will be huge (a given tree will have a number of subtrees that is exponential in its size). Because of this we would like design algorithms whose computational complexity does not depend on  . Representations of this kind have been studied extensively by Bod [2]. However, the work in [2] involves training and decoding algorithms that depend computationally on the num ber of subtrees involved. The parameter estimation techniques described in [2] do not correspond to maximum-likelihood estimation or a discriminative criterion: see [11] for discussion. The methods we propose show that the score for a parse can be calculated in polynomial time in spite of an exponentially large number of subtrees, and that efficient parameter estimation techniques exist which optimize discriminative criteria that have been well-studied theoretically. Goodman [9] gives an ingenious conversion of the model in [2] to an equivalent PCFG whose number of rules is linear in the size of the training data, thus solving many of the computational issues. An exact implementation of Bod?s parsing method is still infeasible, but Goodman gives an approximation that can be implemented efficiently. However, the method still suffers from the lack of justification of the parameter estimation techniques. The key to our efficient use of this high dimensional representation is the definition of an appropriate kernel. We begin by examining the inner product between  two trees   and   under this representation,        #   #   . To compute we first define the set of nodes in trees   and   as   and   respectively. We define the indicator function     to be  if sub-tree  is seen rooted at node  and 0 otherwise. It follows that #            and            . The first step to efficient computation of the inner product is the following property (which can be proved with some simple algebra): # # # #    "!$#&% '(%  )*! ! !2#43 6573 6589:! ! 65-=>5- + -,/.0 + 1,/. + 9,/.0 +  ,;.2$< where we define ?        @     A     . Next, we note that ?        can be computed in polynomial time, due to the following recursive definition: B B If the productions at   and  are different ?         DC . If the productions at   ?           . and    are the same, and   and   are pre-terminals, then In training, a parameter is explicitly estimated for each sub-tree. In searching for the best parse, calculating the score for a parse in principle requires summing over an exponential number of deriva underlying a tree, and in practice is approximated using Monte-Carlo techniques. tions Pre-terminals are nodes directly above words in the surface string, for example the N, V, and D B Else if the productions at   and  are the same and   and  are not pre-terminals,    ?        $"?      !    $    where      is the number of children of   in the tree; because the productions at   / are the same, we have           . The  ?th child-node of   is    $  . To see that this recursive definition is correct, note that ?        simply counts the number of common subtrees that are found rooted at both   and   . The first two cases are trivially correct. The last, recursive, definition follows because a common subtree for   and  can be formed by taking the production at   / , together with a choice at each child of simply taking the non-terminal at that child, or any one of the common sub-trees at that child.                 $$ possible choices at the  ?th child. (Note Thus there are  "?    that a similar recursion is described by Goodman [9], Goodman?s application being the conversion of Bod?s model [2] to an equivalent PCFG.) It is clear from the identity  #           ?        , and the recursive definition of ?      , that  #       can be calculated in         time: the matrix of ?      values can be filled in, then summed. This can be a pessimistic estimate of the runtime. A more useful characterization is that it runs in time linear in the number of members            such that the productions at   and   are the same. In our data we have found a typically linear number of nodes with identical productions, so that most values of ? are 0, and the running time is close to linear in the size of the trees. This recursive kernel structure, where a kernel between two objects is defined in terms of kernels between its parts is quite a general idea. Haussler [10] goes into some detail describing which construction operations are valid in this context, i.e. which operations maintain the essential Mercer conditions. This paper and previous work by Lodhi et al. [12] examining the application of convolution kernels to strings provide some evidence that convolution kernels may provide an extremely useful tool for applying modern machine learning techniques to highly structured objects. The key idea here is that one may take a structured object and split it up into parts. If one can construct kernels over the parts then one can combine these into a kernel over the whole object. Clearly, this idea can be extended recursively so that one only needs to construct kernels over the ?atomic? parts of a structured object. The recursive combination of the kernels over parts of an object retains information regarding the structure of that object. Several issues remain with  the kernel we describe over trees and convolution kernels in general. First, the value of   $   will depend greatly on the size of the trees      .                       # $   One may normalize the kernel by using which also satisfies the essential Mercer conditions. Second, the value of the kernel when applied to two copies of the same tree can be extremely large (in our experiments on the order of  C! ) while the value of the kernel between two different trees is typically much smaller (in our experiments the typical pairwise comparison is of order 100). By analogy with a Gaussian kernel we say that the kernel is very peaked. If one constructs a model which is a linear combination of trees, as one would with an SVM [6] or the perceptron, the output will be dominated by the most similar tree and so the model will behave like a nearest neighbor rule. There are several possible solutions to this problem. Following Haussler [10] we may radialize the kernel, however, it is not always clear that the result is still a valid kernel. Radializing did not appear to help in our experiments. These problems motivate two simple modifications to the tree kernel. Since there will be many more tree fragments of larger size ? say depth four versus depth three ? and symbols in Figure 2. consequently less training data, it makes sense to downweight the contribution of larger tree fragments to the kernel. The first method for doing this is to simply restrict the depth of the tree fragments we consider. The second method is to scale the relative importance of , tree fragments with their size. This can be achieved by introducing a parameter C and modifying the base case and recursive case of the definitions of ? to be respectively     and ?       ?         "?      !   $$          #  $    , where This corresponds to a modified kernel      #      is the number of rules in the  ?th fragment. This kernel downweights the contribution of tree fragments exponentially with their size.   It is straightforward to design similar kernels for tagging problems (see figure 1) and for another common structure found in NLP, dependency structures. See [5] for details. In the tagging kernel, the implicit feature representation tracks all features consisting of a subsequence of state labels, each with or without an underlying word. For example, the paired sequence Lou/SP Gerstner/CP is/N chairman/N of/N IBM/SC would include features such as SP CP , SP Gerstner/CP N , SP CP is/N N of/N and so on.         3 Linear Models for Parsing and Tagging This section formalizes the use of kernels for parsing and tagging problems. The method is derived by the transformation from ranking problems to a margin-based classification problem in [8]. It is also related to the Markov Random Field methods for parsing suggested in [13], and the boosting methods for parsing in [4]. We consider the following set-up: B      Training data is a set of example input/output pairs. In parsing we would have training examples  where each is a sentence and each is the correct tree for that sentence. B    We assume some way of enumerating a set of candidates for a particular sentence. We use   to denote the ?th candidate for the  ?th sentence in training data, and           to denote the set of candidates for . B           Without loss of generality we take  to be the correct parse for (i.e.,    ). B Each candidate  is represented by a feature vector    in the space  . The param . We then define the ?ranking score? of each eters of the model are also a vector   example as    . This score is interpreted as an indication of the plausibility of the      . candidate. The output of the model on a training or test example is  !  !   "$#&%(')"$* + $, ! !  When considering approaches to training the parameter vector , note that a ranking function that correctly ranked the correct parse above all competing candidates would satisfy     4C  . It is simple to modify the Perceptron the conditions $     and Support Vector Machine algorithms to treat this problem. For example, the SVM optimization problem (hard margin version) is to find the which minimizes     subject to  the constraints             . Rather than explicitly calculating , the perceptron algorithm and Support Vector Machines can be formulated as a search ! ! ! - /.  - :.  10 32 42 576 9! 8 ;5 <2 =2 5>6 ! ? This can be achieved using a modified dynamic programming table where 65  =>5  =@  stores 5)-=(58 of depth @ or less. The recursive <definition of can the number of common subtrees at nodes < be modified appropriately. A A context-free grammar ? perhaps taken straight from the training examples ? is one way of enumerating candidates. 5 Another choice is to use a hand-crafted grammar (such as the LFG grammar in [13]) or to take the most probable parses from an existing probabilistic parser (as in [4]).  - - :.  - Define:                              Initialization: Set dual parameters   DC For                If        do nothing, Else       - 10 - 6 Figure 3: The perceptron algorithm for ranking problems. Depth Score Improvement 1 2     3      4     5 6         !  Table 1: Score shows how the parse score varies with the maximum depth of sub-tree considered by the perceptron. Improvement is the relative reduction in error in comparison to the PCFG, which scored 74%. The numbers reported are the mean and standard deviation over the 10 development sets. for ?dual parameters?   which determine the optimal weights ! 8  !    ! 8      -  :.   $ (1)  (we use      as shorthand for     ). It follows that the score of a parse can be calculated using the dual parameters, and inner products between feature vectors, without having to explicitly deal with feature or parameter vectors in the space : ! 8   !            .      For example, see figure 3 for the perceptron algorithm applied to this problem. 4 Experimental Results To demonstrate the utility of convolution kernels for natural language we applied our tree kernel to the problem of parsing the Penn treebank ATIS corpus [14]. We split the treebank randomly into a training set of size 800, a development set of size 200 and a test set of size 336. This was done 10 different ways to obtain statistically significant results. A PCFG was trained on the training set, and a beam search was used to give a set of parses, with PCFG probabilities, for each of the sentences. We applied a variant of the voted perceptron algorithm [7], which is a more robust version of the original perceptron algorithm with performance similar to that of SVMs. The voted perceptron can be kernelized in the same way that SVMs can but it can be considerably more computationally efficient. We generated a ranking problem by having the PCFG generate its top 100 candidate parse trees for each sentence. The voted perceptron was applied, using the tree kernel described previously, to this re-ranking problem. It was trained on 20 trees selected randomly from the top 100 for each sentence and had to choose the best candidate from the top 100 on the test set. We tested the sensitivity to two parameter settings: first, the maximum depth of sub-tree examined, and second, the scaling factor used to down-weight deeper trees. For each value of the parameters we trained on the training set and tested on the development set. We report the results averaged over the development sets in Tables 1 and 2. We report a parse score which combines precision and recall. Define to be the number of correctly placed constituents in the  ?th test tree, " to be the number of constituents Scale Score Imp. 0.1     0.2     0.3    0.4     0.5     0.6     0.7      0.8     0.9   !  Table 2: Score shows how the parse score varies with the scaling factor for deeper sub-trees is varied. Imp. is the relative reduction in error in comparison to the PCFG, which scored 74%. The numbers reported are the mean and standard deviation over the 10 development sets. proposed, and  to be the number of constistuents in the true parse tree. A constituent is defined by a non-terminal label and its span. The score is then    !  C C      "   6 The precision and recall on the  ?th parse are /" and / respectively. The score is then the average precision recall, weighted by the size of the trees  . We also give relative improvements over the PCFG scores. If the PCFG score is and the perceptron score is ,     $ C C   , i.e., the relative reduction in error. the relative improvement is  C C . . We finally used the development set for cross-validation to choose the best parameter settings for each split. We used the best parameter settings (on the development sets) for each split to train on both the training and development sets, then tested on the test set. This gave  C  with the best choice of maximum depth and a score a relative goodness score of of  C  with the best choice of scaling factor. The PCFG scored  on the test data. All of these results were obtained by running the perceptron through the training data only once. As has been noted previously by Freund and Schapire [7], the voted perceptron often obtains better results when run multiple times through the training data. Running through the data twice with a maximum depth of 3 yielded a relative goodness score of    , while using a larger number of iterations did not improve the results significantly. In summary we observe that in these simple experiments the voted perceptron and an appropriate convolution kernel can obtain promising results. However there are other methods which perform considerably better than a PCFG for NLP parsing ? see [3] for an overview ? future work will investigate whether the kernels in this paper give performance gains over these methods. 5 A Compressed Representation When used with algorithms such as the perceptron, convolution kernels may be even more computationally attractive than the traditional radial basis or polynomial kernels. The linear combination of parse trees constructed by the perceptron algorithm can be viewed as a weighted forest. One may then search for subtrees in this weighted forest that occur more than once. Given a linear combination of two trees    "   which contain a common subtree, we may construct a smaller weighted acyclic graph, in which the common subtree occurs only once and has weight   . This process may be repeated until an arbitrary linear combination of trees is collapsed into a weighted acyclic graph in which no subtree occurs more than once. The perceptron may now be evaluated on a new tree by a straightforward generalization of the tree kernel to weighted acyclic graphs of the form produced by this procedure. Given the nature of our data ? the parse trees have a high branching factor, the words are chosen from a dictionary that is relatively small in comparison to the size of the training data, and are drawn from a very skewed distribution, and the ancestors of leaves are part of speech tags ? there are a relatively small number of subtrees in the lower levels of the parse trees that occur frequently and make up the majority of the data. It appears that the approach we have described above should save a considerable amount of computation. This is something we intend to explore further in future work. 6 Conclusions In this paper we described how convolution kernels can be used to apply standard kernel based algorithms to problems in natural language. Tree structures are ubiquitous in natural language problems and we illustrated the approach by constructing a convolution kernel over tree structures. The problem of parsing English sentences provides an appealing example domain and our experiments demonstrate the effectiveness of kernel-based approaches to these problems. Convolution kernels combined with such techniques as kernel PCA and spectral clustering may provide a computationally attractive approach to many other problems in natural language processing. Unfortunately, we are unable to expand on the many potential applications in this short note, however, many of these issues are spelled out in a longer Technical Report [5]. References [1] Aizerman, M., Braverman, E., and Rozonoer, L. (1964). Theoretical Foundations of the Potential Function Method in Pattern Recognition Learning. Automation and Remote Control, 25:821?837. [2] Bod, R. (1998). Beyond Grammar: An Experience-Based Theory of Language. CSLI Publications/Cambridge University Press. [3] Charniak, E. (1997). Statistical techniques for natural language parsing. In AI Magazine, Vol. 18, No. 4. [4] Collins, M. (2000). Discriminative Reranking for Natural Language Parsing. Proceedings of the Seventeenth International Conference on Machine Learning. San Francisco: Morgan Kaufmann. [5] Collins, M. and Duffy, N. (2001). Parsing with a Single Neuron: Convolution Kernels for Natural Language Problems. Technical report UCSC-CRL-01-01, University of California at Santa Cruz. [6] Cortes, C. and Vapnik, V. (1995). Support?Vector Networks. Machine Learning, 20(3):273?297. [7] Freund, Y. and Schapire, R. (1999). Large Margin Classification using the Perceptron Algorithm. In Machine Learning, 37(3):277?296. [8] Freund, Y., Iyer, R.,Schapire, R.E., & Singer, Y. (1998). An efficient boosting algorithm for combining preferences. In Machine Learning: Proceedings of the Fifteenth International Conference. San Francisco: Morgan Kaufmann. [9] Goodman, J. (1996). Efficient algorithms for parsing the DOP model. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP 96), pages 143-152. [10] Haussler, D. (1999). Convolution Kernels on Discrete Structures. Technical report, University of Santa Cruz. [11] Johnson, M. The DOP estimation method is biased and inconsistent. To appear in Computational Linguistics. [12] Lodhi, H., Christianini, N., Shawe-Taylor, J., and Watkins, C. (2001). Text Classification using String Kernels. To appear in Advances in Neural Information Processing Systems 13, MIT Press. [13] Johnson, M., Geman, S., Canon, S., Chi, S., & Riezler, S. (1999). Estimators for stochastic ?unification-based? grammars. In Proceedings of the 37th Annual Meeting of the Association for Computational Linguistics. San Francisco: Morgan Kaufmann. [14] Marcus, M., Santorini, B., & Marcinkiewicz, M. (1993). Building a large annotated corpus of english: The Penn treebank. Computational Linguistics, 19, 313-330. [15] Scholkopf, B., Smola, A.,and Muller, K.-R. (1999). Kernel principal component analysis. In B. Scholkopf, C. J. C. Burges, and A. J. Smola, editors, Advances in Kernel Methods ? SV Learning, pages 327-352. MIT Press, Cambridge, MA. [16] Watkins, C. (2000). Dynamic alignment kernels. In A.J. Smola, P.L. Bartlett, B. Schlkopf, and D. Schuurmans, editors, Advances in Large Margin Classifiers, pages 39-50, MIT Press.
2089 |@word version:2 polynomial:3 lodhi:2 recursively:1 carry:1 reduction:3 contains:2 att:1 fragment:11 score:20 charniak:1 existing:1 com:1 must:1 parsing:19 cruz:3 generative:1 selected:1 leaf:1 reranking:1 short:1 num:1 characterization:1 boosting:2 cse:1 node:7 provides:1 preference:1 ucsc:2 constructed:1 scholkopf:2 shorthand:1 combine:2 pairwise:1 theoretically:1 tagging:6 frequently:1 terminal:5 chi:1 company:1 param:1 considering:2 begin:2 underlying:4 unrelated:1 kind:1 interpreted:1 string:8 minimizes:1 transformation:1 nj:1 formalizes:1 runtime:1 classifier:1 control:1 penn:2 appear:4 modify:1 treat:1 riezler:1 twice:1 initialization:1 studied:2 examined:1 hmms:2 pcfgs:4 statistically:1 averaged:1 seventeenth:1 atomic:1 recursive:9 practice:1 procedure:1 empirical:1 significantly:1 pre:3 word:3 radial:1 spite:1 cannot:1 close:1 context:4 applying:2 collapsed:1 restriction:1 optimize:1 equivalent:2 go:1 straightforward:2 identifying:1 occurences:3 rule:12 haussler:4 estimator:1 searching:1 justification:1 rozonoer:1 mcollins:1 construction:1 parser:1 magazine:1 exact:1 programming:1 us:1 approximated:1 recognition:1 geman:1 observed:1 capture:2 remote:1 substantial:1 complexity:1 dynamic:2 motivate:1 depend:3 solving:1 trained:3 algebra:1 basis:1 jersey:1 various:2 represented:3 train:1 describe:6 monte:1 sc:3 whose:2 quite:1 widely:1 plausible:1 larger:5 say:2 csli:1 otherwise:1 compressed:1 grammar:10 sequence:5 indication:1 propose:1 product:5 combining:1 subgraph:1 bod:4 normalize:1 constituent:3 spelled:1 object:10 tions:1 help:1 nearest:1 strong:1 implemented:1 involves:2 concentrate:1 correct:5 annotated:1 modifying:1 stochastic:2 require:3 generalization:1 marcinkiewicz:1 pessimistic:1 probable:1 considered:1 mapping:2 dictionary:1 estimation:7 label:2 sensitive:2 tool:1 weighted:6 mit:3 clearly:1 gaussian:1 always:1 modified:3 rather:1 publication:1 derived:1 emission:1 improvement:4 prevalent:1 likelihood:2 greatly:1 sense:1 typically:3 entire:1 hidden:8 kernelized:1 ancestor:1 expand:1 issue:4 classification:4 dual:3 proposes:1 development:8 noun:1 summed:1 fairly:1 field:1 construct:4 once:4 having:2 biology:1 identical:1 park:1 imp:2 peaked:1 future:2 np:12 others:1 report:5 modern:1 randomly:2 consisting:1 maintain:1 attempt:1 huge:1 highly:1 investigate:1 braverman:1 evaluation:1 alignment:1 subtrees:8 partial:1 unification:1 experience:1 tree:68 filled:1 taylor:1 re:1 theoretical:1 instance:1 retains:1 goodness:2 applicability:1 introducing:1 deviation:2 examining:2 johnson:2 reported:2 nigel:1 dependency:2 varies:2 sv:1 considerably:3 combined:1 person:2 international:2 sensitivity:1 probabilistic:2 decoding:1 michael:1 together:1 ambiguity:2 choose:2 emnlp:1 potential:2 includes:1 automation:1 satisfy:1 explicitly:3 ranking:6 lab:1 doing:1 start:2 contribution:2 formed:1 voted:6 kaufmann:3 efficiently:1 correspond:1 vp:2 conceptually:1 eters:1 produced:1 schlkopf:1 carlo:1 apple:5 straight:1 suffers:1 definition:7 pp:1 frequency:1 involved:1 gain:1 proved:1 recall:3 knowledge:1 ubiquitous:1 appears:1 higher:1 done:2 evaluated:1 generality:1 implicit:1 smola:3 until:1 hand:1 parse:21 lack:1 perhaps:1 building:1 contain:1 true:1 excluded:1 illustrated:1 deal:2 attractive:2 skewed:1 branching:1 covering:1 rooted:2 noted:1 criterion:3 demonstrate:2 cp:6 common:8 overview:1 attached:1 exponentially:2 extend:1 association:1 atis:3 refer:1 significant:1 cambridge:2 ai:1 trivially:1 neatly:1 language:14 shawe:1 had:1 dot:2 longer:1 surface:1 base:1 something:1 optimizing:1 store:1 continue:1 meeting:1 muller:1 seen:1 morgan:3 canon:1 determine:1 multiple:1 technical:3 calculation:2 downweights:1 plausibility:1 cross:1 paired:1 variant:1 regression:1 essentially:1 fifteenth:1 iteration:1 kernel:63 represent:2 achieved:2 beam:1 else:2 goodman:5 appropriately:1 biased:1 subject:1 member:1 inconsistent:1 effectiveness:1 structural:2 split:4 independence:1 gave:1 competing:2 restrict:1 inner:3 idea:3 regarding:1 avenue:1 enumerating:3 whether:1 pca:2 utility:1 bartlett:1 speech:2 useful:3 santa:3 involve:3 clear:2 amount:2 extensively:1 svms:2 deriva:1 generate:1 schapire:3 exist:1 estimated:2 track:3 correctly:2 discrete:5 vol:1 key:4 four:1 drawn:1 christianini:1 graph:4 convert:2 run:2 named:1 reasonable:1 scaling:3 yielded:1 annual:1 occur:4 constraint:1 tag:2 dominated:1 extremely:2 span:1 downweight:1 relatively:2 department:1 structured:4 combination:5 describes:1 ate:1 remain:1 smaller:2 appealing:1 modification:1 taken:1 computationally:5 previously:2 describing:1 count:5 mechanism:2 singer:1 operation:3 apply:1 observe:1 appropriate:2 spectral:1 save:1 alternative:2 original:2 top:3 running:4 nlp:12 include:1 clustering:1 linguistics:3 calculating:2 giving:1 intend:1 ingenious:1 occurs:2 chairman:6 traditional:1 lou:6 unable:1 entity:2 hmm:1 majority:1 marcus:1 modeled:2 unfortunately:1 design:2 implementation:1 perform:1 allowing:1 conversion:2 convolution:14 neuron:1 markov:2 behave:1 extended:1 santorini:1 dc:2 varied:1 verb:1 arbitrary:1 introduced:2 pair:2 sentence:10 california:2 learned:2 beyond:1 suggested:1 aizerman:1 pattern:1 natural:13 ranked:2 indicator:1 recursion:1 representing:1 improve:1 text:1 relative:10 freund:3 loss:1 par:3 dop:2 analogy:1 versus:1 acyclic:3 validation:1 foundation:1 mercer:3 principle:1 treebank:3 editor:2 share:1 ibm:6 production:9 preposition:1 summary:1 placed:1 last:1 free:3 copy:1 infeasible:1 english:2 allow:1 deeper:2 perceptron:21 ber:1 neighbor:1 burges:1 taking:2 boundary:1 calculated:3 depth:9 transition:1 valid:2 rich:1 san:3 obtains:1 implicitly:2 corpus:4 summing:1 francisco:3 discriminative:4 subsequence:1 search:3 table:4 promising:1 learn:1 nature:1 robust:1 forest:2 schuurmans:1 gerstner:7 constructing:1 domain:2 sp:6 did:2 whole:1 scored:3 nothing:1 child:6 repeated:1 crafted:1 precision:3 sub:12 exponential:2 candidate:9 watkins:3 down:1 symbol:1 disregarding:1 svm:2 cortes:1 evidence:1 essential:2 vapnik:1 pcfg:13 importance:1 subtree:4 iyer:1 duffy:2 margin:4 simply:3 explore:1 corresponds:1 satisfies:1 ma:1 identity:1 formulated:2 viewed:1 consequently:1 jeff:1 replace:1 crl:1 feasible:1 hard:1 considerable:1 included:1 infinite:1 typical:2 principal:2 experimental:3 support:4 collins:3 tested:3
1,195
209
Unsupervised Learning in Neurodynamics Unsupervised Learning in Neurodynamics Using the Phase Velocity Field Approach Michail Zak Nikzad Toornarian Center for Space Microelectronics Technology Jet Propulsion Laboratory California Institute of Technology Pasadena, CA 91109 ABSTRACT A new concept for unsupervised learning based upon examples introduced to the neural network is proposed. Each example is considered as an interpolation node of the velocity field in the phase space. The velocities at these nodes are selected such that all the streamlines converge to an attracting set imbedded in the subspace occupied by the cluster of examples. The synaptic interconnections are found from learning procedure providing selected field. The theory is illustrated by examples. This paper is devoted to development of a new concept for unsupervised learning based upon examples introduced to an artificial neural network. The neural network is considered as an adaptive nonlinear dissipative dynamical system described by the following coupled differential equations: N Ui + K,Ui = L 11j g( Uj ) + Ii i=I,2, ... ,N (I) j=1 in which U is an N-dimensional vector, function of time, representing the neuron activity, T is a constant matrix whose elements represent synaptic interconnections between the neurons, 9 is a monotonic nonlinear function, Ii is the constant exterior input to each neuron, and K, is a positive constant . 583 584 Zak and Toomarian Let us consider a pattern vector u represented by its end point in an n-dimensional phase space, and suppose that this pattern is introduced to the neural net in the form of a set of vectors - examples u Ck ), k 1,2 ... K (Fig. 1). The difference between these examples which represent the same pattern can be caused not only by noisy measurements, but also by the invariance of the pattern to some changes in the vector coordinates (for instance, to translations, rotations etc.). If the set of the points u Ck ) is sufficiently dense, it can be considered as a finite-dimensional approximation of some subspace OCl). = Now the goal of this study is formulated as following: find the synaptic interconnections 7ij and the input to the network h such that any trajectory which is originated inside of OCl) will be entrapped there. In such a performance the subspace OCl) practically plays the role of the basin of attraction to the original pattern U. However, the position of the attractor itself is not known in advance: the neural net has to create it based upon the introduced representative examples. Moreover, in general the attractor is not necessarily static: it can be periodic, or even chaotic. The achievement of the goal formulated above would allow one to incorporate into a neural net a set of attractors representing the corresponding clusters of patterns, where each cluster is imbedded into the basin of its attractor. Any new pattern introduced to such a neural net will be attracted to the "closest" attractor. Hence, the neural net would learn by examples to perform content-addressable memory and pattern recognition. A A \ \ ~- Fig. 1: Two-Dimensional Vectors as Examples, uk, and Formation of Clusters O. Unsupervised Learning in Neurodynamics Our approach is based upon the utilization of the original clusters of the example points u O:) as interpolation nodes of the velocity field in the phase space. The assignment of a certain velocity to an example point imposes a corresponding constraint upon the synaptic interconnections Tij and the input Ii via Eq. (1). After these unknowns are found, the velocity field in the phase space is determined by Eq. (1). Hence, the main problem is to assign velocities at the point examples such that the required dynamical behavior of the trajectories formulated above is provided. One possibility for the velocity selection based upon the geometrical center approach was analyzed by M. Zak, (1989). In this paper a "gravitational attraction" approach to the same problem will be introduced and discussed. Suppose that each example-point u(k) is attracted to all the other points u(k')(k' =j:. k) such that its velocity is found by the same rule as a gravitational force: v~k) = Vo , in which Vo u~k') K - u~k) ?; [2:1 (u?') _ u??)2]3/2 (2) =1 Ir'?Ir is a constant scale coefficient. Actual velocities at the same points are defined by Eq. (1) rearranged as: N u~k) = 2: 7ijg( u~,,) - uod - IC( u~k) - Uoi) j=l i= 1,2, ... ,N k=1,2, ... ,J{ (3) The objective is to find synaptic interconnections Tij and center of gravity Uoi such that they minimize the distance between the assigned velocity (Eq. 2) and actual calculated velocities (Eq. 3). Introducing the energy: (4) one can find Tij and Uoi from the condition: E-min i.e., as the static attractor of the dynamical system: ? uoi = ? T.. ? 2 -(k 8E -- (5a) 8E 87ij (5b) 8U oi 2 --(k-- ') - in which (k is a time scale parameter for learning. By appropriate selection of this parameter the convergence of the dynamical system can be considerably improved (J. Barhen, S. Gulati, and M. Zak, 1989). 585 586 Zak and Toomarian Obviously, the static attractor of Eqs. (5) is unique. As follows from Eq. (3) GU~k) (k) GU j = dg~k) Iij (k)' dU j (i i:- j) (6) d (Ie> Since g(u) is a monotonic function, sgn.f.m is constant which in turn implies that dU j GU~k) sgn -W = Gu. const (i i:- j) (7) 1 Applying this result to the boundary of the cluster one concludes that the velocity at the boundary is directed inside of the cluster (Fig. 2). For numerical illustration of the new learning concept developed above, we select 6 points in the two dimensional space, (i.e., two neurons) which constructs two separated clusters (Fig. 3, points 1-3 and 16-18 (three points are the minimum to form a cluster in two dimensional space?. Coordinates of the points in Fig. 3 are given in Table 1. The assigned velocity vf calculated based on Eq. 2 and Vo 0.04 are shown in dotted line. For a random initialization of Tij and Uoi, the energy decreases sharply from an initial value of 10.608 to less than 0.04 in about 400 iterations and at about 2000 iterations the final value of 0.0328 has been achieved, (Fig. 4). To carry out numerical integration of the differential equations, first order Euler numerical scheme with time step of 0.01 has been used. In this simulation the scale parameter a 2 was kept constant and set to one. By substituting the calculated Iij and Uoi into Eq. (3) for point uk, (k = 1,2,3,16,17,18), one will obtain the calculated velocities at these points (shown as dashed lines in Fig. 3). As one may notice, the assigned and calculated velocities are not exactly the same. However, this small difference between the velocities are of no importance as long as the calculated velocities are directed toward the interior of the cluster. This directional difference of the velocities is one of the reasons that the energy did not vanish. The other reason is the difference in the value of these velocities, which is of no importance either, based on the concept developed. = Fig. 2: Velocities at Boundaries are directed Toward Inside of the Cluster. Unsupervised Learning in Neurodynamics In order to show that for different initial conditions, Eq. 3 will converge to an attractor which is inside one of the two clusters, this equation was started from different points (4-15,19-29). In all points, the equation converges to either (0.709,0.0) or (-0.709,0.0). However, the line x in this case is the dividing line, and all the points on this line will converge to u o . = ? The decay coefficient", and the gain of the hyperbolic tangent were chosen to be 1. However, during the course of this simulation it was observed that the system is very sensitive to these parameters as well as v o , which calls for further study in this area. 15 29 14 4 20 9 Fig. 3:. Cluster 1 (1-3) and Cluster 2 (16-19). Calculated Velocity (- -) ? Assigned Velocity ( .. ) ? Activation Dynamics initiated at different points. 7 587 588 Zak and Thomarian Table 1. - Coordinate of Points in point X Y point X 1 0.50 0.00 16 -0.50 2 1.00 0.25 17 -1.00 3 1.00 -0.25 18 -1.00 4 1.25 0.25 19 -1.25 5 1.25 -0.25 20 -1.25 6 1.00 0.50 21 -1.00 7 1.00 -0.50 22 -1.00 8 0.75 0.50 23 -0.75 9 0.75 -0.50 24 -0.75 10 0.50 0.25 25 -0.50 11 0.50 -0.25 26 -0.50 12 0.25 0.10 27 -0.25 13 0.25 -0.10 28 -0.25 14 0.02 1.00 29 -0.02 15 0.00 1.00 Figure 4. Y 0.00 0.25 0.25 0.25 -0.25 0.50 -0.50 0.50 -0.50 -0.25 -0.25 0.10 -0.10 1.00 \0 0 ? 1"""'4 ? ~ ~ ~ ~ Z C"1 ? I.I"t ~ ? o ?? ..??? ?..??? ? :\..........................:::: ....~ ....~ ....= .....""'. ...------,.-----~ o 100 200 300 ITERATIONS Fig 4: Profile of Neuromorphic Energy over Time Iterations Acknowledgement This research was carried out at the Center for Space Microelectronic Technology, Jet Propulsion Laboratory, California Institute of Technology. Support for the work came from Agencies of the U.S. Department of Defense, including the Innovative Science and Technology Office of the Strategic Defense Initiative Organization and the Office of the Basic Energy Sciences of the US Dept. of Energy, through an agreement with the National Aeronautics and Space Administration. Unsupervised Learning in Neurodynamics References M. Zak (1989), "Unsupervised Learning in Neurondynamics Using Example Interaction Approach", Appl. Math. Letters, Vol. 2, No.3, pp. 381- 286. J. Barhen, S. Gulati, M. Zak (1989), "Neural Learning of Constrained nonlinear Transformations", IEEE Computer, Vol. 22(6), pp. 67-76. 589
209 |@word dividing:1 uj:1 concept:4 implies:1 hence:2 assigned:4 objective:1 laboratory:2 simulation:2 illustrated:1 imbedded:2 sgn:2 during:1 subspace:3 distance:1 carry:1 assign:1 initial:2 propulsion:2 vo:3 toward:2 reason:2 gravitational:2 practically:1 sufficiently:1 considered:3 ic:1 activation:1 geometrical:1 illustration:1 attracted:2 providing:1 rotation:1 numerical:3 substituting:1 discussed:1 selected:2 unknown:1 perform:1 measurement:1 neuron:4 sensitive:1 zak:8 finite:1 create:1 math:1 node:3 incorporate:1 ck:2 occupied:1 attracting:1 etc:1 aeronautics:1 differential:2 office:2 initiative:1 closest:1 introduced:6 required:1 inside:4 certain:1 california:2 came:1 behavior:1 minimum:1 dynamical:4 pattern:8 michail:1 actual:2 pasadena:1 converge:3 provided:1 dashed:1 moreover:1 toomarian:2 ii:3 memory:1 including:1 force:1 development:1 jet:2 constrained:1 integration:1 developed:2 long:1 dept:1 field:5 construct:1 transformation:1 representing:2 scheme:1 technology:5 unsupervised:8 basic:1 started:1 concludes:1 gravity:1 exactly:1 carried:1 coupled:1 iteration:4 uk:2 utilization:1 represent:2 achieved:1 acknowledgement:1 gulati:2 dg:1 tangent:1 positive:1 national:1 phase:5 attractor:8 initiated:1 interpolation:2 organization:1 possibility:1 initialization:1 dissipative:1 basin:2 imposes:1 appl:1 analyzed:1 barhen:2 call:1 translation:1 devoted:1 course:1 directed:3 unique:1 allow:1 chaotic:1 institute:2 procedure:1 addressable:1 administration:1 area:1 hyperbolic:1 defense:2 boundary:3 calculated:7 instance:1 adaptive:1 interior:1 selection:2 assignment:1 neuromorphic:1 applying:1 strategic:1 introducing:1 tij:4 euler:1 center:4 rearranged:1 periodic:1 considerably:1 rule:1 attraction:2 ijg:1 ocl:3 notice:1 dotted:1 ie:1 table:2 neurodynamics:5 learn:1 ca:1 coordinate:3 exterior:1 vol:2 suppose:2 play:1 du:2 necessarily:1 agreement:1 did:1 velocity:23 element:1 recognition:1 kept:1 dense:1 main:1 profile:1 observed:1 role:1 letter:1 fig:10 representative:1 coefficient:2 uoi:6 streamlines:1 caused:1 iij:2 decrease:1 position:1 originated:1 vf:1 agency:1 ui:2 microelectronic:1 vanish:1 dynamic:1 activity:1 minimize:1 oi:1 ir:2 constraint:1 sharply:1 upon:6 decay:1 gu:4 microelectronics:1 directional:1 represented:1 min:1 innovative:1 importance:2 trajectory:2 separated:1 department:1 artificial:1 formation:1 synaptic:5 whose:1 energy:6 pp:2 interconnection:5 static:3 monotonic:2 noisy:1 itself:1 final:1 gain:1 obviously:1 equation:4 net:5 turn:1 goal:2 formulated:3 interaction:1 end:1 content:1 change:1 determined:1 improved:1 appropriate:1 uod:1 invariance:1 achievement:1 entrapped:1 convergence:1 cluster:14 select:1 original:2 nonlinear:3 converges:1 support:1 const:1 ij:2 eq:10
1,196
2,090
A Variational Approach to Learning Curves D?orthe Malzahn Manfred Opper Neural Computing Research Group School of Engineering and Applied Science Aston University, Birmingham B4 7ET, United Kingdom. [malzahnd,opperm]@aston.ac.uk Abstract We combine the replica approach from statistical physics with a variational approach to analyze learning curves analytically. We apply the method to Gaussian process regression. As a main result we derive approximative relations between empirical error measures, the generalization error and the posterior variance. 1 Introduction Approximate expressions for generalization errors for finite dimensional statistical data models can be often obtained in the large data limit using asymptotic expansions. Such methods can yield approximate relations for empirical and true errors which can be used to assess the quality of the trained model see e.g. [1]. Unfortunately, such an approximation scheme does not seem to be easily applicable to popular non-parametric models like Gaussian process (GP) models and Support Vector Machines (SVMs). We apply the replica approach of statistical physics to asses the average case learning performance of these kernel machines. So far, the tools of statistical physics have been successfully applied to a variety of learning problems [2]. However, this elegant method suffers from the drawback that data averages can be performed exactly only under very idealistic assumptions on the data distribution in the ?thermodynamic? limit of infinite data space dimension. We try to overcome these limitations by combining the replica method with a variational approximation. For Bayesian models, our method allows us to express useful data averaged a-posteriori expectations by means of an approximate measure. The derivation of this measure requires no assumptions about the data density and no assumptions about the input dimension. The main focus of this article is Gaussian process regression where we demonstrate the various strengths of the presented method. It solves some of the problems stated at the end of our previous NIPS paper [3] which was based on a simpler somewhat unmotivated truncation of a cumulant expansion. For Gaussian process models we show that our method does not only give explicit approximations for generalization errors but also of their sample fluctuations. Furthermore, we show how to compute corrections to our theory and demonstrate the possibility of deriving approximate universal relations between average empirical and true errors which might be of practical interest. An earlier version of our approach, which was still restricted to the assumption of idealized data distributions appeared in [4]. 2 Setup and Notation We assume that a class of elementary predictors (neural networks, regressors etc.) is given by functions . In a Bayesian formulation, we have a prior distribution over this class of functions .  Assuming  that a set of observations      is conditionally independent    given inputs     , we assign a likelihood term of the form    to each observation. Posterior expectations (denoted by angular brackets) of any functional !#" $ are expressed in the form  ) + ,.!#" $ ,  / 23 5464 * (1)    10   * + where the partition function  normalizes the posterior and denotes % !#" $ &5the 89 expectation with respect to the prior. We are interested in computing averages 7 of posterior ' "   <$  ;  #  expectations over different drawings of training data sets : % !#" $ &(' were all data examples are independently generated from the same distribution. In the next section we will show how to derive a measure which enables us to compute analytically approximate combined data and posterior averages. 3 A Grand-Canonical Approach We utilize the statistical mechanics approach to>the of learning. Our aim is to =1? * analysis 8@9  which serves as a generating compute the so-called averaged ?free energy? 7 * function for suitable data averages of posterior expectations. The partition function  is  ,  /  . 3  4   (2) 10  =A? * 8B9 =1? * 8B9C'C=AD1EGFHJILK.M NPO Q.F SUR TWV  we use the replica trick 7 , To perform K * F 8B9 the average 7  where 7  is computed for integer X and the continuation is performed at the end [5]. We obtain F  * FYZ[ ' * F 8B9\' + F^]_G` ,  /  a . 54Jbc1de fghi (3) 7    a0  + F * where  ' + denotes the expectation over the replicated prior measure. Eq.(3) can be transformed into a simpler form by introducing the ?grand canonical? partiF kl tion function j F kl 'n / m Zs r * F Z>t' + F   vu F  (4) 0YIpo q  with the Hamiltonian F uwFx'y ` ,  / l a  z4Jbc1de fg (5)   a0  oq g ~F R evaluates all8 c1X de freplicas   of the predictor at the same data The  density   point  } is taken with respect to the true data density ?  and theo|{expectation 7W   . F?kl The ?grand canonical? partition functionc j g represents a ?poissonized? version k of the original model with fluctuating number of examples. The ?chemical potential? deterZ?' K.M N? R k?'?=A?Z mines the expected value of which yields simply for X???? . For K Z q sufficiently large , we can replace theq sum in Eq. (4) by its dominating term =A? * FYZ>t??=A? FYkll??Zs=1??Z? U?Z?k ) j (6) j c g c g thereby neglecting relative fluctuations. We recover the original (canonical) free energy as  K M NQ F R  ?  K6M N? R F M N  . K K 4 Variational Approximation FYkl For most interesting cases, the partition function j can not be computed in closed u F form for given X . Hence, by a different u F I we use a variational approach to approximate tractable Hamiltonian . It is easy to write down the firstu terms F  u inF I an expansion of the ?grand canonical? free energy with respect to the difference >=A? j FYkl ' >=1? + F % &zI o {} I  R ? % uwFY?u F 5& I ? )  % u#F??u F I   &5I; % uwFY?u F I & I   (7) The brackets denote averages with respect to the effective measure which is induced  R and acts in the space of replicated variables. As is well known, by the prior and o;{.} terms in Eq.(7) present an upper bound [6] to >=A? j FYkl . Although the first two leading differentiating the bound with respect to X will u F I usually not preserve the inequality, we still is a sensible thing to do [7]. expect 1 that an optimization with respect to 4.1 Variational Equations The grand-canonical ensemble was chosen  such that Eq.(5)u can Z  |?as an" integral  $  F ' berewritten a over a local quantity in the input variable , i.e. in the form with F s " a .<$  '    ,  / l  a  z4  ?    a0   (8) We will now specialize to Gaussian priors over , for which a local quadratic expression u FI '  | / a ) a . a    .l? /   a % a  a . &I (9) . is a suitable . trial Hamiltonian, leading to Gaussian averages  . The functions  a and  a are variational parameters to be optimized. It is important to have an explicit de  pendence on the input variable in order to take a non uniform input density into account. To perform the variation of the first > two Eq.(7) =A? terms u F I  ? we% uwnote F that u F I &zthe I locality of Eq.(8) + F inzv makes the ?variational free energy? an explicit function  of the first two local moments  a . ' % a .  .&5I  Hence, a straightforward variation yields Z % s   &zI ' .  a   a  Z a . ' % a  &5I % ? . &zI '  . a (10)   a  (11) To extend the variational solutions to non-integer values ?'  of . X , we assumeathat .?for ' all.  the  a optimal parameters are replica symmetric, ie. as well as for    '  ' I|    notation for a . and  a  . and  aa . We also use a corresponding   1 Guided by the success of the method in physical applications, for instance in polymer physics. 4.2 Interpretation of u FI Note, that our approach uwI is not equivalent to a variational approximation of the original posterior. In contrast, contains the full information of the statistics of the training data. R in order to compute approximate We can use the distribution induced by the prior and . { } o combined data and posterior averages. As an example, we first consider the expected local %  .&  % .&  8 9  . ' 7 . Following the algebra of the replica posterior variance method (see [5]) this is approximated within the variational replica approach as   ' |F =AD1H E I  % a   & I  % a   . & I ' I .t   . (12) Second, we consider the noisy local mean square prediction error of the posterior mean  ' %  &   '  .t   8@9 predictor which is given by   7  . In this case      ' '  F =AD1H E I  % a .   &5I?      % a .&5I    U   ??  (  (13) We can also calculate fluctuations with respect to the data average, for example 7  .t   | B  @8 9 ' F =AD1H E I     e   (  e  B @  0   0a  a I   (14) 5 Regression with Gaussian Processes  ' 23 ?   , where is This statistical model assumes that data   are generated as  Gaussian white  noise   t' with .  58 { . The prior lover  functions t'    has.zero   mean and + variance covariance  7 . Hence, we have  . Using the   definitions Eqs.(12,13), we get  % ? " a .<$ &5I 'y . 1= ? ?? | .?  ?   ) X  .    ?       ?    (15) { which yields the set of variational equations (11). They become    particularly easy when the and the input distribution regression model uses a translationally invariant kernel  is homogeneous in a finite interval. The variational equations (11) can then be solved in terms of the eigenvalues of the Gaussian process kernel. [8, 9] studied learning curves for Gaussian process regression which are not only averaged  over the data but also over the data generating process using a Gaussian process prior on  . Applying these averages of.[9]  to. our  theory  and .adapting ?  the notation  simply replaces  .  ?     by   . while in Eq.(15) the term  5.1 Learning Curves and Fluctuations Practical situations differ from this ?typical case? analysis. The data generating process is unknown but assumedto be fixed. The resulting learning curve is then conditioned on this particular ?teacher? . The left panel of Fig.1 shows an example. Displayed are the mean square prediction error  (circle and solid line) and its sample fluctuations   (error bars)  with respect to the data average (cross and broken line). The target was a random but fixed realization    p' "from ! az Gaussian    $process #  &% '  prior  ' ' with a periodic Radial Basis Function kernel   , ? ) . We keep the example   ' simple,  ' ? ?e.g) . the Gaussian process regression model used the same kernel and noise)( 8 The inputs are one dimensional, independent and uniformly distributed { 7 ? ) . Symbols represent simulation data. A typical property of our theory (lines) is that it becomes very accurate for sufficiently large number of example data. 0 0 Theory: Lines Simulation: Symbols Correction of Free Energy Generalization Error ?, Fluctuation ?? 10 ?1 10 ?2 10 ? ?3 10 ?1 ? =0.25 ?1 ?1 ? =0.01 ?1 ? =0.0001 ?2 ?3 ?? ?4 10 0 100 50 150 Number m of Example Data ?4 0 200 20 40 60 80 Number m of Example Data 100 Figure 1: Gaussian  ( process8 regression using a periodic Radial Basis Function kernel, input 7 ? ) , and homogeneous dimension d=1,  density. Left: Generalization error   '   ' input and fluctuations  for data noise ? ? ) . Right: Correction of the free energy. { Symbols: We subtracted the first two contributions to Eq.(7) from the true value of the free energy. The latter was obtained by simulations. Lines show the third contribution of Eq.(7).  decreases from top to bottom. All y-data was set equal The value of the noise variance { to zero. 5.2 Corrections to the Variational Approximation It is a strength of our method that the quality of the variational approximation Eq.(7) can be characterized and systematically improved. In this paper, we restrict ourself to a characterization and  consider the case where all  -data is set equal to zero. Since the posterior variance is independent of the data this is still an interesting model from which the posterior variance can be estimated. We consider the third term in the expansion to the free energy Eq.(7). It is a correction to the variational free energy and evaluates to d ed I F & I  'y[) 7  I .  I B  ; B58   Z I         `  b d ed    ??   (16)   {   & I  . Eq.(16)   is shown by lines in the right panel of Fig.1 for different values of the model noise . We considered a homo{ geneous input density, the input dimension is one and the regression model uses a periodic   I FA= DAH E I   ) % u F  u F  & I  % u F  u X Z       ? `=A? ,  4Jb d e d ? )     l?^      {     ' =1DAE FH I % a . a   p a .  with    RBF kernel. The symbols in Fig.1 show the difference between the true value of the free energy which is obtained by simulations and the first two terms of Eq.(7). The correction term is found to be qualitatively accurate and emphasizes a discrepancy between free energy and the first two terms of the expansion Eq.(7) for a medium amount of example data. The calculated learning curves inherit this behaviour. 5.3 Universal Relations We can relate the training error  and the empirical posterior variance     ' ) ` / Z A 0          b 9     ' ) ` / Z 1 0 ;   b   9  (17) 0.8 0.8 0.6 2 2 0.6 2 Theory 1d, periodic 2d, periodic 3d, periodic 2 [??(x,y)/(?? (x)+1) ](x,y) 1 [?? (x)/(?? (x)+1)]x 1 0.4 Theory d=1, periodic d=2, periodic d=3, periodic d=2+2, non-periodic 0.2 0 0 0.2 0.4 ??T 2 0.8 0.6 0.4 0.2 0 0 1 0.2 0.4 ??T 0.6 0.8 1 Figure 2: Illustration of relation Eq.(19) (left) and Eq.(20) (right). All error measures are  scaled with . Symbols show simulation results for Radial Basis Function (RBF) regression x' )   dimensions (square, circle, diamond). and a homogeneous input distribution in The RBF kernel was periodic. Additionally, the left figure shows an ? example were the ' inputs lie on a quasi two-dimensional manifold which is embedded in dimensions (cross). In this case the RBF kernel was non-periodic. ?=A? * 89 '   ?    to the free energy   7     . Using Eqs.(6,7) and the stationarity of  the grand-canonical free energy with respect to the variational parameters we obtain the following relation | 7 >=1? *     8B9 ?yZ  % ? " a  $ &5I  X |  (18) We use the fact that the posterior variance is independent of the  -data and simply estimate it from the model where all  -data is set equal to zero. In this case, Eq.(18) yields   '    ? ) ?^    .   (19) . which relates at  the empirical posterior variance  to the local posterior variance test inputs . Similarly, we can derive an expression for the training error   by using Eqs.(15,18) in combination with Eq.(19)   ' | ??       ? ?      .  ) (20) It is interesting to note, that the relations (19,20) contain no assumptions about the data generating process. They hold in general for Gaussian process models with a Gaussian likelihood. An illustration of Eqs.(19,20) is given by Fig.2 for the example of Gaussian process regression with a Radial Basis Function kernel. In the left panel of  Fig.2, learning  starts in the upper right corner as the rescaled empirical posterior variance  is initially one and decreases with increasing number of example data. For the right panel of Fig.2, learning starts in the lower left corner. The rescaled training error   on the noisy data set is initially zero and increases to one with increasing number of example data. The theory (line) holds for a sufficiently large number of example data and its accuracy increases with the input dimension. Eqs.(19,20) can also be tested on real data. For common benchmark sets such as Abalone and Boston Housing data we find that Eqs.(19,20) hold well even for small and medium sizes of the training data set. 6 Outlook One may question if our approximate universal relations are of any practical use as, for example, the relation between training error and generalization error involves also the un . known posterior variance . Nevertheless, this relation could be useful for cases, where a large number of data inputs without output labels are available. Since for regression, the posterior variance is independent of the output labels, we could use these extra input points   to estimate . The application of our technique to more complicated # .models .  is possible and technically ) in Eq.(1) and further rescalby more involved. For example, replacing     t'      % o { ing the kernel  of the Gaussian process prior gives a model for hard      . The condition margin Support Vector Machine Classification with SVM kernel  of maximum margin classification will be ensured by the limes ? . Of particular interest is the computation of empirical estimators that can be used in practice for model selection as well as the calculation of fluctuations (error bars) for such estimators. A prominent example is an efficient approximate leave-one-out estimator for SVMs. Work on these issues is in progress. Acknowledgement We would like to thank Peter Sollich for may inspiring discussions. The work was supported by EPSRC grant GR/M81601. References [1] N. Murata, S. Yoshizawa, S. Amari, IEEE Transactions on Neural Networks 5, p. 865-872, (1994). [2] A. Engel, C. Van den Broeck, Statistical Mechanics of Learning, Cambridge University Press (2001). [3] D. Malzahn, M. Opper, Neural Information Processing Systems 13, p. 273, T. K. Leen, T. G. Dietterich and V. Tresp, eds., MIT Press, Cambridge MA (2001). [4] D. Malzahn, M. Opper, Lecture Notes in Computer Science 2130, p. 271, G. Dorffner, H. Bischof and K. Hornik, eds., Springer, Berlin (2001). [5] M. M?ezard, G. Parisi, M. Virasoro, Spin Glass Theory and Beyond, World Scientific, Singapore, (1987). [6] R. P. Feynman and A. R. Hibbs, Quantum mechanics and path integrals, Mc GrawHill Inc., (1965). [7] T. Garel, H. Orland, Europhys. Lett. 6, p. 307 (1988). [8] P. Sollich, Neural Information Processing Systems 11, p. 344, M. S. Kearns, S. A. Solla and D. A. Cohn, eds., MIT Press, Cambridge MA (1999). [9] P. Sollich, Neural Information Processing Systems 14, T. G. Dietterich, S. Becker, Z. Ghahramani, eds., MIT Press (2002).
2090 |@word trial:1 version:2 simulation:5 covariance:1 thereby:1 outlook:1 solid:1 moment:1 contains:1 united:1 partition:4 enables:1 hamiltonian:3 manfred:1 characterization:1 simpler:2 become:1 specialize:1 combine:1 expected:2 mechanic:3 increasing:2 becomes:1 notation:3 panel:4 medium:2 z:2 orland:1 act:1 exactly:1 ensured:1 scaled:1 uk:1 grant:1 engineering:1 local:6 limit:2 fluctuation:8 path:1 might:1 studied:1 averaged:3 practical:3 vu:1 practice:1 universal:3 empirical:7 adapting:1 radial:4 get:1 selection:1 applying:1 equivalent:1 straightforward:1 independently:1 estimator:3 deriving:1 variation:2 target:1 homogeneous:3 approximative:1 us:2 hjilk:1 trick:1 approximated:1 particularly:1 bottom:1 epsrc:1 solved:1 calculate:1 solla:1 decrease:2 rescaled:2 broken:1 mine:1 ezard:1 trained:1 algebra:1 technically:1 basis:4 easily:1 various:1 derivation:1 effective:1 europhys:1 dominating:1 drawing:1 amari:1 statistic:1 gp:1 noisy:2 housing:1 eigenvalue:1 parisi:1 combining:1 realization:1 opperm:1 generating:4 leave:1 derive:3 ac:1 school:1 progress:1 eq:25 solves:1 involves:1 differ:1 guided:1 drawback:1 behaviour:1 assign:1 generalization:6 polymer:1 elementary:1 correction:6 hold:3 sufficiently:3 b9:4 considered:1 idealistic:1 birmingham:1 applicable:1 label:2 engel:1 successfully:1 tool:1 mit:3 gaussian:18 aim:1 focus:1 likelihood:2 contrast:1 glass:1 posteriori:1 a0:3 initially:2 relation:10 transformed:1 quasi:1 interested:1 theq:1 classification:2 issue:1 denoted:1 equal:3 represents:1 discrepancy:1 jb:2 preserve:1 uwi:1 translationally:1 malzahnd:1 stationarity:1 interest:2 possibility:1 homo:1 bracket:2 pendence:1 accurate:2 integral:2 neglecting:1 circle:2 dae:1 virasoro:1 instance:1 earlier:1 introducing:1 predictor:3 uniform:1 gr:1 teacher:1 periodic:12 broeck:1 combined:2 density:6 grand:6 ie:1 physic:4 m81601:1 corner:2 leading:2 account:1 potential:1 de:1 inc:1 idealized:1 performed:2 try:1 tion:1 closed:1 analyze:1 start:2 recover:1 complicated:1 contribution:2 ass:2 square:3 spin:1 accuracy:1 variance:13 ensemble:1 yield:5 murata:1 bayesian:2 emphasizes:1 mc:1 suffers:1 ed:6 definition:1 evaluates:2 energy:13 npo:1 involved:1 yoshizawa:1 popular:1 improved:1 formulation:1 leen:1 furthermore:1 angular:1 replacing:1 cohn:1 quality:2 scientific:1 dietterich:2 contain:1 true:5 analytically:2 hence:3 chemical:1 symmetric:1 orthe:1 twv:1 white:1 conditionally:1 abalone:1 prominent:1 demonstrate:2 variational:17 fi:2 common:1 functional:1 physical:1 b4:1 extend:1 interpretation:1 cambridge:3 z4:1 similarly:1 etc:1 posterior:19 inf:1 inequality:1 success:1 somewhat:1 relates:1 thermodynamic:1 full:1 ing:1 characterized:1 calculation:1 cross:2 prediction:2 regression:11 expectation:7 kernel:12 represent:1 interval:1 extra:1 induced:2 elegant:1 thing:1 oq:1 lover:1 seem:1 integer:2 easy:2 variety:1 zi:3 restrict:1 expression:3 becker:1 peter:1 dorffner:1 hibbs:1 useful:2 amount:1 inspiring:1 svms:2 continuation:1 canonical:7 singapore:1 estimated:1 write:1 express:1 group:1 nevertheless:1 replica:8 utilize:1 sum:1 lime:1 bound:2 hi:1 quadratic:1 replaces:1 strength:2 combination:1 sollich:3 den:1 restricted:1 invariant:1 taken:1 equation:3 tractable:1 feynman:1 end:2 serf:1 available:1 rewritten:1 apply:2 fluctuating:1 subtracted:1 original:3 denotes:2 assumes:1 top:1 ghahramani:1 question:1 quantity:1 parametric:1 thank:1 berlin:1 sensible:1 ourself:1 manifold:1 fy:6 assuming:1 sur:1 illustration:2 kingdom:1 unfortunately:1 setup:1 relate:1 stated:1 unknown:1 perform:2 diamond:1 upper:2 observation:2 benchmark:1 finite:2 displayed:1 situation:1 kl:7 optimized:1 bischof:1 nip:1 malzahn:3 beyond:1 c1d:2 bar:2 usually:1 appeared:1 suitable:2 scheme:1 aston:2 tresp:1 prior:10 acknowledgement:1 asymptotic:1 relative:1 embedded:1 expect:1 lecture:1 interesting:3 limitation:1 article:1 systematically:1 normalizes:1 supported:1 truncation:1 free:13 theo:1 differentiating:1 distributed:1 van:1 curve:6 opper:3 dimension:7 overcome:1 calculated:1 world:1 quantum:1 lett:1 qualitatively:1 regressors:1 replicated:2 far:1 transaction:1 approximate:9 keep:1 assumed:1 un:1 additionally:1 hornik:1 expansion:5 inherit:1 main:2 noise:5 fig:6 explicit:3 lie:1 third:2 down:1 symbol:5 svm:1 conditioned:1 margin:2 boston:1 locality:1 simply:3 expressed:1 springer:1 ma:2 rbf:4 replace:1 hard:1 infinite:1 typical:2 uniformly:1 kearns:1 called:1 support:2 latter:1 cumulant:1 tested:1
1,197
2,091
Fast and Robust Classification using Asymmetric AdaBoost and a Detector Cascade Paul Viola and Michael Jones Mistubishi Electric Research Lab Cambridge, MA viola@merl.com and mjones@merl.com Abstract This paper develops a new approach for extremely fast detection in domains where the distribution of positive and negative examples is highly skewed (e.g. face detection or database retrieval). In such domains a cascade of simple classifiers each trained to achieve high detection rates and modest false positive rates can yield a final detector with many desirable features: including high detection rates, very low false positive rates, and fast performance. Achieving extremely high detection rates, rather than low error, is not a task typically addressed by machine learning algorithms. We propose a new variant of AdaBoost as a mechanism for training the simple classifiers used in the cascade. Experimental results in the domain of face detection show the training algorithm yields significant improvements in performance over conventional AdaBoost. The final face detection system can process 15 frames per second, achieves over 90% detection, and a false positive rate of 1 in a 1,000,000. 1 Introduction In many applications fast classification is almost as important as accurate classification. Common examples include robotics, user interfaces, and classification in large databases. In this paper we demonstrate our approach in the domain of low latency, sometimes called ?real-time?, face detection. An extremely fast face detector is a critical component in many applications. User-interfaces can be constructed which detect the presence and number of users. Teleconference systems can automatically devote additional bandwidth to participant?s faces. Video security systems can record facial images of individuals after unauthorized entry. Recently we presented a real-time face detection system which scans video images at 15 frames per second [8] yet achieves detection rates comparable with the best published results (e.g. [7]) 1 Face detection is a scanning process, in which a face classifier is evaluated at every scale and location within each image. Since there are about 50,000 unique scales 1 In order to achieve real-time speeds other systems often resort to skin color filtering in color images or motion filtering in video images. These simple queues are useful but unreliable. In large image databases color and motion are often unavailable. Our system detects faces using only static monochrome information. and locations in a typical image, this amounts to evaluating the face classifier 750,000 times per second. One key contribution of our previous work was the introduction of a classifier cascade. Each stage in this cascade was trained using AdaBoost until the required detection performance was achieved [2]. In this paper we present a new training algorithm designed specifically for a classifier cascade called asymmetric AdaBoost. The algorithm is a generalization of that given in Singer and Shapire [6]. Many of the formal guarantees presented by Singer and Shapire also hold for this new algorithm. The paper concludes with a set of experiments in the domain of face detection demonstrating that asymmetric AdaBoost yields a significant improvement in detection performance over conventional boosting. 2 Classifier Cascade In the machine learning community it is well known that more complex classification functions yield lower training errors yet run the risk of poor generalization. If the main consideration is test set error, structural risk minimization provides a formal mechanism for selecting a classifier with the right balance of complexity and training error [1]. Another significant consideration in classifier design is computational complexity. Since time and error are fundamentally different quantities, no theory can simply select the optimal trade-off. Nevertheless, for many classification functions computation time is directly related to the structural complexity. In this way temporal risk minimization is clearly related to structural risk minimization. This direct analogy breaks down in domains where the distribution over the class labels is highly skewed. For example, in the domain of face detection, there are at most a few dozen faces among the 50,000 sub-windows in an image. Surprisingly in these domains it is often possible to have the best of both worlds: high detection rates and extremely fast classification. The key insight is that while it may be impossible to construct a simple classifier which can achieve a low training/test error, in some cases it is possible to construct a simple classifier with a very low false negative rate. For example, in the domain of face detection, we have constructed an extremely fast classifier with a very low false negative rate (i.e. it almost never misses a face) and a 50% false positive rate. Such a detector might be more accurately called a classification pre-filter: when an image region is labeled ?nonface? then it can be immediately discarded, but when a region is labeled ?face? then further classification effort is required. Such a pre-filter can be used as the first stage in a cascade of classifiers (see Figure 1). In our face detection application (described in more detail in Section 5) the cascade has 38 stages. Even though there are many stages, most are not evaluated for a typical nonface input window since the early stages weed out many non-faces. In fact, over a large test set, the average number of stages evaluated is less than 2. In a cascade, computation time and detection rate of the first few stages is critically important to overall performance. The remainder of the paper describes techniques for training cascade classifiers which are efficient yet effective. 3 Using Boosting to Train the Cascade In general almost any form of classifier can be used to construct a cascade; the key properties are that computation time and the detection rate can be adjusted. Examples include support vector machines, perceptrons, and nearest neighbor classifiers. In the case of an SVM computation time is directly related to the number of support vectors and detection rate is related to the margin threshold [1]. All Sub?windows T 1 F T 2 F T 3 Further Processing F Reject Sub?window Figure 1: Schematic depiction of a detection cascade. A sequence of classifiers are applied to every example. The initial classifier eliminates a large number of negative examples with very little processing. Subsequent stages eliminate additional negatives but require additional computation. Extremely few negative examples remain after several stages. In our system each classifier in the cascade is a single layer perceptron whose input is a set of computationally efficient binary features. The computational cost of each classifier is then simply the number of input features. The detection rate is adjusted by changing the threshold (or bias). Much of the power of our face detection system comes from the very large and varied set of features available. In our experiments over 6,000,000 different binary features were available for inclusion in the final classifiers (see Figure 4 for some example features). The efficiency of each classifier, and hence the efficiency of the cascade, is ensured because a very small number of features are included in the early stages; the first stage has 1 (!) feature, the second stage 5 features, then 20, and then 50. See Section 5 for a brief description of the feature set. The main contribution of this paper is the adaptation of AdaBoost for the task of feature selection and classifier learning. Though it is not widely appreciated, AdaBoost provides a principled and highly efficient mechanism for feature selection[2, 6]. If the set of weak classifiers is simply the set of binary features (this is often called boosting stumps) each round of boosting adds a single feature to the set of current features.  AdaBoost is an iterative process in which each round selects a weak classifier, minimizes:      !   $#&% , which (1)  is the weight on example at round " , Following  (')*the '+ notation of Shapire and Singer,  !, is the target label of the example, is the example, and is a confidence rated binary classifier[6]. After every round the weights are updated as follows: -/. 01 2        3-465879;: <;=>= (2) CBDE5F7>9;: <G@H@ The classifier takes on two possible values and <G@>= ,  <?=A@ where IKJ*L is the weight of the examples given the label which have true label M . These predictions insure that the weights on the next round are balanced: that the relative weights of positive and negative examples one each side of the classification boundary are equal.     Minimizing minimizes the weighted exponential loss round " . Minimizing in each    atwhich round is also a greedy technique for minimizing N is an upper bound on the training error of the strong classifier. It has also been observed that the example weights are directly related to example margin, which leads to a principled argument for AdaBoost?s generalization capabilities [5]. The key advantage of AdaBoost as a feature selection mechanism, over competitors such as the wrapper method [3], is the speed of learning. Given the constraint that the search   , over features is greedy, AdaBoost efficiently selects the feature which minimizes N a surrogate for overall classification error. The entire dependence on previously selected features is efficiently and compactly encoded using the example weights. As a result, the addition of the 100th feature requires no more effort than the selection of the first feature. 2 4 Asymmetric AdaBoost One limitation of AdaBoost arises in the context of skewed example distributions and cascaded classifiers: AdaBoost minimizes a quantity related to classification error; it does not minimize the number of false negatives. Given that the final form of the classifier is a weighted majority of features, the detection and false positive rates are adjustable after training. Unfortunately feature selection proceeds as if classification error were the only goal, and the features selected are not optimal for the task of rejecting negative examples. One naive scheme for ?fixing? AdaBoost is to modify the initial distribution over the training examples. If we hope to minimize false negatives then the weight on positive examples could be increased so that the minimum error criteria will also have very few false negatives. We can formalize this intuitive approach as follows. Recall that AdaBoost is a scheme which minimizes:   $ /  A   (3)  Each term in the summation is bounded above by a simple loss function:  >         0$  '     ) if otherwise   where is the class assigned by the boosted classifier. As a result, minimizing minimizes an upper bound on simple loss. (4)   N  We can introduce a related notion of asymmetric loss:   '     6  (' ) if   (and '    1 E' )   01   .  if and (5)   otherwise where false negatives cost " than false positives. Note that   0  >!* 5879  times 0 . more If we take the bound in Equation 4 and by    "  we obtain a bound on the asymmetric loss: multiply $# both sides     %  .     H       Minimization of this bound can be achieved using AdaBoost by pre-weighting each ex"   >  5879 ample by . The derivation is identical to that of Equation 3. Expanding 2   Given that there are millions of features and thousands of examples, the boosting process requires days of computation. Many other techniques while feasible for smaller problems are likely to be infeasible for this sort of problem. Equation 2 repeatedly for  0 -/. 1  B3. 0 we arrive at,   H  5879   N  # in terms of  " ) (6)   where the second term in the numerator arises because of the initial asymmetric weighting. Noticing that the left hand side must sum to 1 yields the following equality,   $ ! /  A  D  F5 7>9 "  (7)   Therefore AdaBoost minimizes the required bound on asymmetric loss. Unfortunately this naive technique is only somewhat effective. The main reason is AdaBoost?s balanced reweighting scheme. As a result the initially asymmetric example weights are immediately lost. Essentially the AdaBoost process is too greedy. The first classifier selected absorbs the entire effect of the initial asymmetric weights. The remaining rounds are entirely symmetric. We propose a closely related approach that results in the minimization of the same bound, throughout all rounds. Instead of applywhich nevertheless preserves the asymmetric loss  >  5879  ing the necessary asymmetric multiplier at the first round of an round  " .  5F7>9 "  process, the nth root is applied before each round. Referring to Equation 6 we can see the final effect is the same; this preserves the bound on asymmetric loss. But the effect on the training process is quite different. In order to demonstrate this approach we generated an artificial data set and learned strong classifiers containing 4 weak classifiers. The results are shown inFigure 2. In this figure we can see that all but the first weak classifier learned by the naive rule are poor, since they each balance positive and negative errors. The final combination of these classifiers cannot yield high detection rates without introducing many false positives. All the weak classifiers generated by the proposed Asymmetric Adaboost rule are consistent with asymmetric loss and the final strong classifier yields very high detection rates and modest false positive rates.     One simple reinterpretation of this distributed scheme for asymmetric reweighting     ,$  . 5F7>9 is as a reduction in the positive confidence of each weak classifier . This forces each subsequent weak classifier to focus asymmetrically on postive examples. 5 Experiments   We performed two experiments in the domain of frontal face detection to demonstrate the advantages of asymmetric AdaBoost. Experiments follow the general form, though differ in details, from those presented in Viola and Jones [8]. In each round of boosting one of a very large set of binary features are selected. These features, which we call rectangle features, are briefly described in Figure 4. In the first experiment a training and test set containing faces and non-faces of a fixed size were acquired (faces were scaled to a size     pixels). The training set consisted of 1500 face examples and 5000 non-face examples. Test data included 900 faces and 5000 nonfaces. The face examples were manually cropped from a large collection of Web images while the non-face examples were randomly chosen patches from Web images that were known not to contain any faces. Naive asymetric AdaBoost and three parameterizations of Asymmetric AdaBoost were used to train classifiers with 4 features on this data. Figure 3 shows the ROC curves on Figure 2: Two simple examples: positive examples are ?x?, negative ?o? and weak classifiers are linear separators. On the left is the naive asymetric result. The first feature selected is labelled ?1?. Subsequent features attempt to balance positive and negative errors. Notice that no linear combination of the 4 weak classifiers can achieve a low false positive and low false negative rate. On the right is the asymetric boosting result. After learning 4 weak classifier the positives are well modelled and most of the negative are rejected. 0.995 0.99 0.985 0.98 0.975 NAIVE T11-F10 T15-F10 T20-F10 0.97 0.2 0.25 0.3 0.35 0.4 0.45 0.5 0.55 0.6 0.65 0.7 Figure 3: ROC curves for four boosted classifier with 4 features. The first is naive asymmetric boosting. The other three results are for the new asymmetric approach, each using slightly different parameters. The ROC curve has been cropped to show only the region of interest in training a cascaded detector, the high detection rate regime. Notice that that at 99% detection asymmetric Adaboost cuts the false positive by about 20%. This will significantly reduce the work done by later stages in the cascade.                                  B A         C                                    D Figure 4: Left: Example rectangle features shown relative to the enclosing detection window. The sum of the pixels which lie within the white rectangles are subtracted from the sum of pixels in the gray rectangles. A threshold operation is then applied to yield a binary output. Two-rectangle features are shown in (A) and (B). Figure (C) shows a threerectangle feature, and (D) a four-rectangle feature. Right: The first two example feature selected by the boosting process. Notice that the first feature relies on the fact that the horizontal region of the eyes is darker than the horizontal region of the cheeks. The second feature, whose selection is conditioned on the first, acts to distinguish horizontal edges from faces by looking for a strong vertical edge near the nose. test data for the three classifiers. The key result here is that at high detection rates the false positive rate can be reduced significantly. In the second experiment, naive and asymmetric AdaBoost were used to train two different complete cascaded face detectors. Performance of each cascade was determined on a realworld face detection task, which requires scanning of the cascade across a set of large images which contain embedded faces. The cascade training process is complex, and as a result comparing detection results is useful but potentially risky. While the data used to train the two cascades were identical, the performance of earlier stages effects the selection of non-faces used to train later stages. As a result different non-face examples are used to train the corresponding stages for the Naive and Asymmetric results. Layers were added to each of the cascades until the number of false positives was reduced below 100 on a validation set. For normal boosting this occurred with 34 layers. For asymmetric AdaBoost this occurred with 38 layers. Figure 5 shows the ROC curves for the resulting face detectors on the MIT+CMU [4] test set. 3 Careful examination of the ROC curves show that the asymmetric cascade reduces the number of false positives significantly. At a detection rate of 91% the reduction is by a factor of 2. 6 Conclusions We have demonstrated that a cascade classification framework can be used to achieve fast classification, high detection rates, and very low false positive rates. The goal for each classifier in the cascade is not low error, but instead extremely high detection rates and modest false positive rates. If this is achieved, each classifier stage can be used to filter out and discard many negatives. 3 Note: the detection and false positive rates for the simple 40 feature experiment and the more complex cascaded experiment are not directly comparable, since the test sets are quite different. ROC curves for face detector with different boosting algorithms correct detection rate 0.95 0.9 0.85 Asymmetric Boosting Normal Boosting 0.8 0 50 100 150 200 250 300 false positives Figure 5: ROC curves comparing the accuracy of two full face detectors, one trained using normal boosting and the other with asymmetric AdaBoost. Again, the detector trained using asymmetric AdaBoost is more accurate over a wide range of false positive values. Many modern approaches for classification focus entirely on the minimization of errors. Questions of relative loss only arise in the final tuning of the classifier. We propose a new training algorithm called asymmetric AdaBoost which performs learning and efficient feature selection with the fundamental goal of achieving high detection rates. Asymmetric AdaBoost is a simple modification of the ?confidence-rated? boosting approach of Singer and Shapire. Many of their derivations apply to this new approach as well. Experiments have demonstrated that asymmetric AdaBoost can lead to significant improvements both in classification speed and in detection rates. References [1] Corinna Cortes and Vladimir Vapnik. Support-vector networks. Machine Learning, 20, 1995. [2] Yoav Freund and Robert E. Schapire. A decision-theoretic generalization of on-line learning and an application to boosting. In Computational Learning Theory: Eurocolt ?95, pages 23?37. Springer-Verlag, 1995. [3] G. John, R. Kohavi, and K. Pfleger. Irrelevant features and the subset selection problem. In Machine Learning Conference, pages 121?129. Morgan Kaufmann, 1994. [4] H. Rowley, S. Baluja, and T. Kanade. Neural network-based face detection. In IEEE Patt. Anal. Mach. Intell., volume 20, pages 22?38, 1998. [5] R. E. Schapire, Y. Freund, P. Bartlett, and W. S. Lee. Boosting the margin: a new explanation for the effectiveness of voting methods. Ann. Stat., 26(5):1651?1686, 1998. [6] Robert E. Schapire and Yoram Singer. Improved boosting algorithms using confidence-rated predictions. Machine Learning, 37:297?336, 1999. [7] H. Schneiderman and T. Kanade. A statistical method for 3D object detection applied to faces and cars. In Computer Vision and Pattern Recognition, 2000. [8] Paul Viola and Michael J. Jones. Robust real-time object detection. In Proc. of IEEE Workshop on Statistical and Computational Theories of Vision, 2001.
2091 |@word briefly:1 reduction:2 initial:4 wrapper:1 pfleger:1 selecting:1 current:1 com:2 comparing:2 yet:3 must:1 john:1 subsequent:3 designed:1 greedy:3 selected:6 record:1 provides:2 parameterizations:1 boosting:18 location:2 constructed:2 direct:1 absorbs:1 introduce:1 acquired:1 detects:1 eurocolt:1 automatically:1 little:1 window:5 notation:1 insure:1 bounded:1 minimizes:7 guarantee:1 temporal:1 every:3 act:1 voting:1 ensured:1 classifier:49 scaled:1 positive:26 before:1 modify:1 mach:1 might:1 range:1 unique:1 lost:1 cascade:25 reject:1 significantly:3 pre:3 confidence:4 cannot:1 selection:9 risk:4 impossible:1 context:1 conventional:2 demonstrated:2 immediately:2 insight:1 rule:2 notion:1 updated:1 target:1 user:3 recognition:1 asymmetric:31 cut:1 database:3 labeled:2 observed:1 thousand:1 region:5 trade:1 principled:2 balanced:2 complexity:3 rowley:1 trained:4 reinterpretation:1 efficiency:2 compactly:1 derivation:2 train:6 fast:8 effective:2 artificial:1 whose:2 encoded:1 widely:1 quite:2 otherwise:2 final:8 sequence:1 advantage:2 propose:3 remainder:1 adaptation:1 achieve:5 f10:3 description:1 intuitive:1 object:2 stat:1 fixing:1 nearest:1 strong:4 come:1 differ:1 closely:1 correct:1 filter:3 require:1 generalization:4 summation:1 adjusted:2 hold:1 normal:3 achieves:2 early:2 f7:2 proc:1 label:4 weighted:2 minimization:6 hope:1 mit:1 clearly:1 rather:1 boosted:2 monochrome:1 focus:2 improvement:3 detect:1 typically:1 eliminate:1 entire:2 initially:1 selects:2 pixel:3 overall:2 classification:17 among:1 equal:1 construct:3 never:1 manually:1 identical:2 jones:3 develops:1 fundamentally:1 few:4 modern:1 randomly:1 preserve:2 intell:1 individual:1 attempt:1 detection:45 asymetric:3 interest:1 highly:3 multiply:1 accurate:2 edge:2 necessary:1 facial:1 modest:3 merl:2 increased:1 earlier:1 yoav:1 cost:2 introducing:1 entry:1 subset:1 too:1 scanning:2 referring:1 fundamental:1 lee:1 off:1 michael:2 again:1 containing:2 f5:1 resort:1 stump:1 performed:1 break:1 root:1 lab:1 later:2 sort:1 participant:1 capability:1 contribution:2 minimize:2 accuracy:1 kaufmann:1 efficiently:2 yield:8 weak:10 modelled:1 accurately:1 critically:1 rejecting:1 published:1 detector:10 competitor:1 static:1 recall:1 color:3 car:1 formalize:1 day:1 follow:1 adaboost:33 improved:1 evaluated:3 though:3 done:1 rejected:1 stage:17 until:2 hand:1 horizontal:3 web:2 reweighting:2 gray:1 b3:1 effect:4 consisted:1 true:1 multiplier:1 contain:2 hence:1 assigned:1 equality:1 symmetric:1 white:1 round:13 skewed:3 nonface:2 numerator:1 criterion:1 complete:1 demonstrate:3 theoretic:1 performs:1 motion:2 interface:2 image:12 consideration:2 recently:1 common:1 volume:1 million:1 occurred:2 significant:4 cambridge:1 tuning:1 inclusion:1 depiction:1 add:1 irrelevant:1 discard:1 verlag:1 binary:6 morgan:1 minimum:1 additional:3 somewhat:1 full:1 desirable:1 reduces:1 ing:1 retrieval:1 schematic:1 prediction:2 variant:1 vision:2 essentially:1 cmu:1 sometimes:1 robotics:1 achieved:3 addition:1 cropped:2 addressed:1 kohavi:1 eliminates:1 ample:1 effectiveness:1 call:1 structural:3 near:1 presence:1 bandwidth:1 reduce:1 bartlett:1 effort:2 queue:1 repeatedly:1 useful:2 latency:1 amount:1 reduced:2 schapire:3 notice:3 per:3 patt:1 key:5 four:2 demonstrating:1 nevertheless:2 achieving:2 threshold:3 changing:1 rectangle:6 sum:3 run:1 realworld:1 noticing:1 schneiderman:1 arrive:1 almost:3 throughout:1 cheek:1 patch:1 decision:1 comparable:2 entirely:2 layer:4 bound:8 distinguish:1 constraint:1 speed:3 argument:1 extremely:7 combination:2 poor:2 describes:1 remain:1 smaller:1 slightly:1 across:1 modification:1 ikj:1 computationally:1 equation:4 previously:1 mechanism:4 singer:5 nose:1 infigure:1 available:2 operation:1 apply:1 subtracted:1 corinna:1 remaining:1 include:2 t11:1 yoram:1 cbd:1 shapire:4 skin:1 added:1 quantity:2 question:1 dependence:1 surrogate:1 devote:1 majority:1 reason:1 balance:3 minimizing:4 vladimir:1 unfortunately:2 robert:2 potentially:1 negative:18 design:1 enclosing:1 anal:1 adjustable:1 upper:2 vertical:1 postive:1 discarded:1 viola:4 looking:1 frame:2 varied:1 community:1 required:3 security:1 learned:2 proceeds:1 below:1 pattern:1 regime:1 t20:1 including:1 video:3 explanation:1 power:1 critical:1 force:1 examination:1 cascaded:4 nth:1 scheme:4 rated:3 brief:1 eye:1 risky:1 concludes:1 naive:9 relative:3 embedded:1 loss:10 freund:2 limitation:1 filtering:2 analogy:1 validation:1 consistent:1 surprisingly:1 infeasible:1 appreciated:1 formal:2 bias:1 side:3 perceptron:1 neighbor:1 wide:1 face:41 distributed:1 boundary:1 curve:7 evaluating:1 world:1 collection:1 unreliable:1 search:1 iterative:1 kanade:2 robust:2 expanding:1 unavailable:1 complex:3 separator:1 electric:1 domain:10 main:3 paul:2 arise:1 weed:1 roc:7 darker:1 sub:3 exponential:1 lie:1 weighting:2 dozen:1 down:1 svm:1 cortes:1 workshop:1 false:25 vapnik:1 conditioned:1 margin:3 simply:3 likely:1 springer:1 relies:1 ma:1 goal:3 ann:1 careful:1 labelled:1 feasible:1 included:2 typical:2 specifically:1 determined:1 baluja:1 miss:1 called:5 asymmetrically:1 experimental:1 t15:1 perceptrons:1 select:1 support:3 scan:1 arises:2 frontal:1 ex:1
1,198
2,092
On Spectral Clustering: Analysis and an algorithm Andrew Y. Ng CS Division U.C. Berkeley ang@cs.berkeley.edu Michael I. Jordan CS Div. & Dept. of Stat. U.C. Berkeley jordan@cs.berkeley.edu Yair Weiss School of CS & Engr. The Hebrew Univ. yweiss@cs.huji.ac.il Abstract Despite many empirical successes of spectral clustering methodsalgorithms that cluster points using eigenvectors of matrices derived from the data- there are several unresolved issues. First, there are a wide variety of algorithms that use the eigenvectors in slightly different ways. Second, many of these algorithms have no proof that they will actually compute a reasonable clustering. In this paper, we present a simple spectral clustering algorithm that can be implemented using a few lines of Matlab. Using tools from matrix perturbation theory, we analyze the algorithm, and give conditions under which it can be expected to do well. We also show surprisingly good experimental results on a number of challenging clustering problems. 1 Introduction The task of finding good clusters has been the focus of considerable research in machine learning and pattern recognition. For clustering points in Rn-a main application focus of this paper- one standard approach is based on generative models, in which algorithms such as EM are used to learn a mixture density. These approaches suffer from several drawbacks. First, to use parametric density estimators , harsh simplifying assumptions usually need to be made (e.g., that the density of each cluster is Gaussian) . Second, the log likelihood can have many local minima and therefore multiple restarts are required to find a good solution using iterative algorithms. Algorithms such as K-means have similar problems. A promising alternative that has recently emerged in a number of fields is to use spectral methods for clustering. Here, one uses the top eigenvectors of a matrix derived from the distance between points. Such algorithms have been successfully used in many applications including computer vision and VLSI design [5, 1]. But despite their empirical successes, different authors still disagree on exactly which eigenvectors to use and how to derive clusters from them (see [11] for a review). Also, the analysis of these algorithms, which we briefly review below, has tended to focus on simplified algorithms that only use one eigenvector at a time. One line of analysis makes the link to spectral graph partitioning, in which the sec- ond eigenvector of a graph's Laplacian is used to define a semi-optimal cut. Here, the eigenvector is seen as a solving a relaxation of an NP-hard discrete graph partitioning problem [3], and it can be shown that cuts based on the second eigenvector give a guaranteed approximation to the optimal cut [9, 3]. This analysis can be extended to clustering by building a weighted graph in which nodes correspond to datapoints and edges are related to the distance between the points. Since the majority of analyses in spectral graph partitioning appear to deal with partitioning the graph into exactly two parts, these methods are then typically applied recursively to find k clusters (e.g. [9]). Experimentally it has been observed that using more eigenvectors and directly computing a k way partitioning is better (e.g. [5, I]). Here, we build upon the recent work of Weiss [11] and Meila and Shi [6], who analyzed algorithms that use k eigenvectors simultaneously in simple settings. We propose a particular manner to use the k eigenvectors simultaneously, and give conditions under which the algorithm can be expected to do well. 2 Algorithm Given a set of points S = {81' ... ,8 n } in jRl that we want to cluster into k subsets: 1. Form the affinity matrix A E R nx n defined by A ij i # j , and A ii = O. = exp(-Ilsi - sjW/2( 2 ) if 2. Define D to be the diagonal matrix whose (i , i)-element is the sum of A's i-th row, and construct the matrix L = D-l / 2AD-l / 2 . 1 3. Find Xl , X2 , ... , Xk , the k largest eigenvectors of L (chosen to be orthogonal to each other in the case of repeated eigenvalues), and form the matrix X = [XIX2 . . . Xk) E R n xk by stacking the eigenvectors in columns. 4. Form the matrix Y from X by renormalizing each of X's rows to have unit length (i.e. Yij = X ij/CL.j X~)1 / 2). 5. Treating each row of Y as a point in Rk , cluster them into k clusters via K-means or any other algorithm (that attempts to minimize distortion). 6. Finally, assign the original point Si to cluster j if and only if row i of the matrix Y was assigned to cluster j. Here, the scaling parameter a 2 controls how rapidly the affinity Aij falls off with the distance between 8i and 8j, and we will later describe a method for choosing it automatically. We also note that this is only one of a large family of possible algorithms, and later discuss some related methods (e.g., [6]). At first sight, this algorithm seems to make little sense. Since we run K-means in step 5, why not just apply K-means directly to the data? Figure Ie shows an example. The natural clusters in jR2 do not correspond to convex regions, and Kmeans run directly finds the unsatisfactory clustering in Figure li. But once we map the points to jRk (Y 's rows) , they form tight clusters (Figure lh) from which our method obtains the good clustering shown in Figure Ie. We note that the clusters in Figure lh lie at 90 0 to each other relative to the origin (cf. [8]). lReaders familiar with spectral graph theory [3) may be more familiar with the Laplacian 1- L. But as replacing L with 1- L would complicate our later discussion, and only changes the eigenvalues (from Ai to 1 - Ai ) and not the eigenvectors, we instead use L . 3 3.1 Analysis of algorithm Informal discussion: The "ideal" case To understand the algorithm, it is instructive to consider its behavior in the "ideal" case in which all points in different clusters are infinitely far apart. For the sake of discussion, suppose that k = 3, and that the three clusters of sizes n1, n2 and n3 are 8 1 ,82 , and 8 3 (8 = 8 1 U 8 2 U 8 3 , n = n1 +n2 + n3)' To simplify our exposition, also assume that the points in 8 = {Sl,'" ,Sn} are ordered according to which cluster they are in, so that the first n1 points are in cluster 8 1 , the next n2 in 8 2 , etc. We will also use "j E 8/' as a shorthand for s? E 8 i . Moving the clusters "infinitely" far apart corresponds to zeroing all the efements Aij corresponding to points Si and Sj in different clusters. More precisely, define Aij = 0 if Xi and Xj are in different clusters, and Aij = Aij otherwise. Also let t , D , X and Y be defined as in the previous algorithm, but starting with A instead of A. Note that A and t are therefore block-diagonal: A. = [ A(ll) 0 0 A(22) o o o 1; A L = [L(11) 0 o ?(22) (1) A~~ 0 0 o where we have adopted the convention of using parenthesized superscripts to index into subblocks of vectors/matrices, and Lrii) = (D(ii)) - 1/2A(ii) (D(ii)) - 1/2. Here, A(ii) = A(ii) E jRni xni is the matrix of "intra-cluster" affinities for cluster i. For future use, also define d(i) E jRni to be the vector containing D(ii) 's diagonal elements, and dE jRn to contain D's diagonal elements. To construct X, we find t's first k = 3 eigenvectors. Since t is block diagonal, its eigenvalues and eigenvectors are the union of the ei~envalues and eigenvectors of its blocks (the latter padded appropriately with zeros). It is straightforward to show that Lrii) has a strictly positive principal eigenvector xii) E jRni with eigenvalue 1. Also, since A)~) > 0 (j i:- k), the next eigenvalue is strictly less than 1. (See, e.g., [3]). Thus, stacking t 's eigenvectors in columns to obtain X= xi1) [ 0 o xi 2) 0 0 0 0 xi 3) X, we have: 1 E jRnx3. (2) Actually, a subtlety needs to be addressed here. Since 1 is a repeated eigenvalue in t, we could just as easily have picked any other 3 orthogonal vectors spanning the same subspace as X's columns, and defined them to be our first 3 eigenvectors. That is, X could have been replaced by XR for any orthogonal matrix R E jR3X3 (RT R = RRT = 1). Note that this immediately suggests that one use considerable caution in attempting to interpret the individual eigenvectors of L, as the choice of X's columns is arbitrary up to a rotation, and can easily change due to small perturbations to A or even differences in the implementation of the eigensolvers. Instead, what we can reasonably hope to guarantee about the algorithm will be arrived at not by considering the (unstable) individual columns of X, but instead the subspace spanned by the columns of X, which can be considerably more stable. Next, when we renormalize each of X's rows to have unit length, we obtain: y= [ y(l) y(2) jRni xk 0r 00 1 R (3) to denote the i-th subblock of Y. Letting fiji) y(3) where we have used y(i) E 1 [r0 0 0 r denote the j-th row of17(i) , we therefore see that fjY) is the i-th row ofthe orthogonal matrix R. This gives us the following proposition. Proposition 1 Let A's off-diagonal blocks A(i j ) , i =I- j, be zero. Also assume that each cluster Si is connected. 2 Then there exist k orthogonal vectors 1'1, . .. ,1' k (1'; l' j = 1 if i = j, 0 otherwise) so that Y's rows satisfy , (i) ~ ( ) 4 =G for all i = 1, ... ,k, j = 1, ... ,ni. In other words , there are k mutually orthogonal points on the surface of the unit k-sphere around which Y 's rows will cluster. Moreover, these clusters correspond exactly to the true clustering of the original data. 3.2 The general case In the general case, A's off-diagonal blocks are non-zero, but we still hope to recover guarantees similar to Proposition 1. Viewing E = A - A as a perturbation to the "ideal" A that results in A = A+E, we ask: When can we expect the resulting rows of Y to cluster similarly to the rows of Y? Specifically, when will the eigenvectors of L, which we now view as a perturbed version of L, be "close" to those of L? Matrix perturbation theory [10] indicates that the stability of the eigenvectors of a matrix is determined by the eigengap. More precisely, the subspace spanned by L's first 3 eigenvectors will be stable to small changes to L if and only if the eigengap 8 = IA3 - A41, the difference between the 3rd and 4th eigenvalues of L, is large. As discussed previously, the eigenvalues of L is the union of the eigenvalues of D11), D22), and D33), and A3 = 1. Letting Ay) be the j-th largest eigenvalue of Dii), we therefore see that A4 = maxi A~i). Hence, the assumption that IA3 - A41 be large is exactly the assumption that maXi A~i) be bounded away from 1. Assumption AI. There exists 8 > 0 so that, for all i = 1, ... ,k, A~i) :s: 1 - 8. Note that A~i) depends only on Dii), which in turn depends only on A(ii) = A(ii) , the matrix of intra-cluster similarities for cluster Si' The assumption on A~i) has a very natural interpretation in the context of clustering. Informally, it captures the idea that if we want an algorithm to find the clusters Sl, S2 and S3, then we require that each of these sets Si really look like a "tight" cluster. Consider an example in which Sl = S1.1 U S1.2 , where S1.1 and S1.2 are themselves two well separated clusters. Then S = S1.1 U S1.2 U S2 U S3 looks like (at least) four clusters, and it would be unreasonable to expect an algorithm to correctly guess what partition of the four clusters into three subsets we had in mind. This connection between the eigengap and the cohesiveness of the individual clusters can be formalized in a number of ways. Assumption ALl. Define the Cheeger constant [3] of the cluster Si to be _. h(S.) - mmI ~lE I, kIi' I A;;.,') '(.) ~ ,(.)}. . {~ mm lEI d , kli'I where the outer minimum is over all index subsets I there exists 8 > 0 so that (h(Si))2 /2 ~ 8 for all i. 2This condition is satisfied by A.j~) > 0 (j i- k) , which (5) dk ~ {I, ... ,nd. Assume that is true in our case. A standard result in spectral graph theory shows that Assumption Al.l implies Assumption Al. Recall that d)i) = 2:k A)~) characterizes how "well connected" or how "similar" point j is to the other points in the same cluster. The term in the minI{?} characterizes how well (I , I) partitions Si into two subsets, and the minimum over I picks out the best such partition. Specifically, if there is a partition of Si'S points so that the weight of the edges across the partition is small, and so that each of the partitions has moderately large "volume" (sum of dY) 's), then the Cheeger constant will be small. Thus, the assumption that the Cheeger constants h(Si) be large is exactly that the clusters Si be hard to split into two subsets. We can also relate the eigengap to the mixing time of a random walk (as in [6]) defined on the points of a cluster, in which the chance of transitioning from point i to j is proportional to A ij , so that we tend to jump to nearby-points. Assumption Al is equivalent to assuming that, for such a walk defined on the points of any one of the clusters Si , the corresponding transition matrix has second eigenvalue at most 1- 8. The mixing time of a random walk is governed by the second eigenvalue; thus, this assumption is exactly that the walks mix rapidly. Intuitively, this will be true for tight (or at least fairly "well connected") clusters, and untrue if a cluster consists of two well-separated sets of points so that the random walk takes a long time to transition from one half of the cluster to the other. Assumption Al can also be related to the existence of multiple paths between any two points in the same cluster. Assumption A2. There is some fixed fl i l =j:. i 2, we have that > 0, so that for every iI , i2 E {I, ... ,k} , (6) To gain intuition about this, consider the case of two "dense" clusters il and i2 of size O(n) each. Since dj measures how "connected" point j is to other points in the same cluster, it will be dj = O(n) in this case, so the sum, which is over 0(n 2 ) terms , is in turn divided by djdk = O(n 2 ) . Thus, as long as the individual Ajk's are small, the sum will also be small, and the assumption will hold with small fl. Whereas dj measures how connected Sj E Si is to the rest of Si, 2:k:k'itSi Ajk measures how connected Sj is to points in other clusters. The next assumption is that all points must be more connected to points in the same cluster than to points in other clusters; specifically, that the ratio between these two quantities be small. Assumption A3. For some fixed f2 > 0, for every i = 1, ... ,k, j E Si, we have: (7) For intuition about this assumption, again consider the case of densely connected clusters (as we did previously). Here, the quantity in parentheses on the right hand side is 0(1), so this becomes equivalent to demanding that the following ratio be small: (2:k:k'it Si Ajk)/dj = (2: k:k'it Si Ajk)/(2:k:kESi A jk ) = 0(f2) . Assumption A4. There is some constant C > . _ ' (i) ni ' (i ) J - 1, ... ,ni, we have dj ~ (2: k =l dk )/(Cni). ? so that for every i = 1, .. . ,k, This last assumption is a fairly benign one that no points in a cluster be "too much less" connected than other points in the same cluster. Theorem 2 Let assumptions Al, A2, A3 and A4 hold. Set f = Jk(k - l)fl + kE~. If 0 > (2 + V2}::, then there exist k orthogonal vectors rl, . .. , rk (rF r j = I if i = j, o otherwise) so that Y's rows satisfy (8) Thus, the rows of Y will form tight clusters around k well-separated points (at 90 0 from each other) on the surface of the k-sphere according to their "true" cluster Si. 4 Experiments To test our algorithm, we applied it to seven clustering problems. Note that whereas was previously described as a human-specified parameter, the analysis also suggests a particularly simple way of choosing it automatically: For the right (J2, Theorem 2 predicts that the rows of Y will form k "tight" clusters on the surface of the k-sphere. Thus, we simply search over (J2 , and pick the value that, after clustering Y 's rows, gives the tightest (smallest distortion) clusters. K-means in Step 5 of the algorithm was also inexpensively initialized using the prior knowledge that the clusters are about 90 0 apart. 3 The results of our algorithm are shown in Figure l a-g. Giving the algorithm only the coordinates of the points and k, the different clusters found are shown in the Figure via the different symbols (and colors, where available). The results are surprisingly good: Even for clusters that do not form convex regions or that are not cleanly separated (such as in Figure 19) , the algorithm reliably finds clusterings consistent with what a human would have chosen. (J2 We note that there are other, related algorithms that can give good results on a subset of these problems, but we are aware of no equally simple algorithm that can give results comparable to these. For example, we noted earlier how K-means easily fails when clusters do not correspond to convex regions (Figure Ii). Another alternative may be a simple "connected components" algorithm that, for a threshold T, draws an edge between points Si and Sj whenever Iisi - sjl12 :s: T, and takes the resulting connected components to be the clusters. Here, T is a parameter that can (say) be optimized to obtain the desired number of clusters k. The result of this algorithm on the threecircles-j oined dataset with k = 3 is shown in Figure lj. One of the "clusters" it found consists of a singleton point at (1.5,2). It is clear that this method is very non-robust. We also compare our method to the algorithm of Meila and Shi [6] (see Figure lk). Their method is similar to ours, except for the seemingly cosmetic difference that they normalize A's rows to sum to I and use its eigenvectors instead of L 's, and do not renormalize the rows of X to unit length. A refinement of our analysis suggests that this method might be susceptible to bad clusterings when the degree to which different clusters are connected (L: j d;il) varies substantially across clusters. 3 Briefiy, we let the first cluster centroid be a randomly chosen row of Y , and then repeatedly choose as the next centroid the row of Y that is closest to being 90? from all the centroids (formally, from the worst-case centroid) already picked. The resulting K-means was run only once (no restarts) to give the results presented. K-means with the more conventional random initialization and a small number of restarts also gave identical results. In contrast, our implementation of Meila and Shi 's algorithm used 2000 restarts. flips,8clusten o o (a) (b) (c) th reeci~es-joiJ\ed,2c1ust 8fS squigg les, 4 clusteNl (d) Ih reecirdes_joined,3clusters (e) RowsoJYOittered , rarKlomlysubsa m pled) lorlW<lCirc~ (i) ~nea r.dballs , 3 dus\efs(Meil a and Shi algor1lhm) o o q, o~ 00 0 o~ ~& 0 o o 0 &~llO ~~ o 0 o (j) lWo circles, 2 cluSle<S (K_means) (h) (g) threecircles-joined, 3 clusters(conoecled """'l""'enlS) (f) (k) flips, 6 cluste<s (Kannan elal ,aigor;thm) N (I) Figure 1: Clustering examples, with clusters indicated by different symbols (and colors, where available). (a-g) Results from our algorithm, where the only parameter varied across runs was k. (h) Rows of Y (jittered, subsampled) for twocircles dataset . (i) K-means. (j) A "connected components" algorithm. (k) Meila and Shi algorithm. (1) Kannan et al. Spectral Algorithm I. (See text.) 5 Discussion There are some intriguing similarities between spectral clustering methods and Kernel peA, which has been empirically observed to perform clustering [7, 2]. The main difference between the first steps of our algorithm and Kernel PCA with a Gaussian kernel is the normalization of A (to form L) and X. These normalizations do improve the performance of the algorithm, but it is also straightforward to extend our analysis to prove conditions under which Kernel PCA will indeed give clustering. While different in detail , Kannan et al. [4] give an analysis of spectral clustering that also makes use of matrix perturbation theory, for the case of an affinity matrix with row sums equal to one. They also present a clustering algorithm based on k singular vectors , one that differs from ours in that it identifies clusters with individual singular vectors. In our experiments, that algorithm very frequently gave poor results (e.g., Figure 11). Acknowledgments We thank Marina Meila for helpful conversations about this work. We also thank Alice Zheng for helpful comments. A. Ng is supported by a Microsoft Research fellowship. This work was also supported by a grant from Intel Corporation, NSF grant IIS-9988642, and ONR MURI N00014-00-1-0637. References [1] C. Alpert, A. Kahng, and S. Yao. Spectral partitioning: The more eigenvectors, the better. Discrete Applied Math , 90:3- 26, 1999. [2] N. Christianini, J. Shawe-Taylor, and J. Kandola. Spectral kernel methods for clustering. In Neural Information Processing Systems 14, 2002. [3] F. Chung. Spectral Graph Theory. Number 92 in CBMS Regional Conference Series in Mathematics. American Mathematical Society, 1997. [4] R. Kannan, S. Vempala, and A. Yetta. On clusterings- good, bad and spectral. In Proceedings of the 41st Annual Symposium on Foundations of Computer Science, 2000. [5] J. Malik, S. Belongie, T. Leung, and J. Shi. Contour and texture analysis for image segmentation. In Perceptual Organization for Artificial Vision Systems. Kluwer, 2000. [6] M. Meila and J. Shi. Learning segmentation by random walks. In N eural Information Processing Systems 13, 200l. [7] B. Scholkopf, A. Smola, and K.-R Miiller. Nonlinear component analysis as a kernel eigenvalue problem. N eural Computation, 10:1299- 1319, 1998. [8] G. Scott and H. Longuet-Higgins. Feature grouping by relocalisation of eigenvectors of the proximity m atrix. In Proc. British Machine Vision Conference, 1990. [9] D. Spielman and S. Teng. Spectral partitioning works: Planar graphs and finite element meshes. In Proceedings of the 37th Annual Symposium on Foundations of Computer Science, 1996. [10] G. W. Stewart and J.-G. Sun. Matrix Perturbation Th eory. Academic Press, 1990. [11] Y . Weiss. Segmentation using eigenvectors: A unifying view. In International Conf erence on Computer Vision, 1999.
2092 |@word version:1 briefly:1 seems:1 nd:1 cleanly:1 llo:1 simplifying:1 pick:2 atrix:1 recursively:1 eigensolvers:1 series:1 ours:2 si:19 intriguing:1 must:1 mesh:1 partition:6 benign:1 treating:1 rrt:1 generative:1 half:1 guess:1 xk:4 math:1 node:1 mathematical:1 symposium:2 scholkopf:1 shorthand:1 consists:2 prove:1 manner:1 indeed:1 expected:2 behavior:1 themselves:1 frequently:1 automatically:2 little:1 considering:1 becomes:1 moreover:1 bounded:1 nea:1 what:3 substantially:1 eigenvector:5 caution:1 finding:1 corporation:1 untrue:1 guarantee:2 berkeley:4 every:3 exactly:6 partitioning:7 unit:4 control:1 grant:2 appear:1 positive:1 local:1 despite:2 meil:1 path:1 might:1 initialization:1 suggests:3 challenging:1 alice:1 acknowledgment:1 ond:1 union:2 block:5 differs:1 xr:1 empirical:2 erence:1 word:1 close:1 context:1 equivalent:2 map:1 conventional:1 shi:7 straightforward:2 starting:1 convex:3 d22:1 ke:1 formalized:1 immediately:1 estimator:1 higgins:1 spanned:2 datapoints:1 stability:1 coordinate:1 suppose:1 us:1 origin:1 element:4 recognition:1 jk:2 particularly:1 cut:3 predicts:1 muri:1 observed:2 capture:1 worst:1 region:3 connected:13 sun:1 cheeger:3 intuition:2 moderately:1 iisi:1 efs:1 engr:1 solving:1 tight:5 upon:1 division:1 f2:2 easily:3 univ:1 separated:4 describe:1 artificial:1 choosing:2 whose:1 emerged:1 distortion:2 say:1 otherwise:3 superscript:1 seemingly:1 eigenvalue:13 propose:1 unresolved:1 j2:3 rapidly:2 mixing:2 normalize:1 cluster:69 renormalizing:1 derive:1 andrew:1 ac:1 stat:1 ij:3 school:1 lwo:1 implemented:1 c:6 implies:1 convention:1 drawback:1 pea:1 human:2 viewing:1 dii:2 require:1 kii:1 assign:1 yweiss:1 really:1 proposition:3 yij:1 strictly:2 mm:1 hold:2 around:2 proximity:1 exp:1 a2:2 smallest:1 proc:1 largest:2 successfully:1 tool:1 weighted:1 hope:2 gaussian:2 sight:1 derived:2 focus:3 unsatisfactory:1 likelihood:1 indicates:1 cosmetic:1 contrast:1 centroid:4 sense:1 helpful:2 leung:1 typically:1 lj:1 vlsi:1 issue:1 fairly:2 field:1 construct:2 once:2 aware:1 ng:2 equal:1 identical:1 look:2 future:1 np:1 simplify:1 few:1 randomly:1 simultaneously:2 densely:1 kandola:1 individual:5 familiar:2 replaced:1 subsampled:1 n1:3 microsoft:1 attempt:1 inexpensively:1 organization:1 intra:2 zheng:1 mixture:1 analyzed:1 d11:1 xni:1 edge:3 lh:2 orthogonal:7 taylor:1 walk:6 initialized:1 desired:1 renormalize:2 circle:1 column:6 earlier:1 stewart:1 stacking:2 subset:6 too:1 perturbed:1 varies:1 jittered:1 considerably:1 st:1 density:3 international:1 huji:1 ie:2 off:3 xi1:1 michael:1 yao:1 again:1 satisfied:1 containing:1 choose:1 conf:1 american:1 chung:1 li:1 jrk:1 de:1 singleton:1 sec:1 satisfy:2 ad:1 depends:2 later:3 view:2 picked:2 analyze:1 characterizes:2 recover:1 minimize:1 il:3 ni:3 who:1 correspond:4 ofthe:1 cni:1 mmi:1 tended:1 whenever:1 complicate:1 ed:1 proof:1 gain:1 dataset:2 ask:1 recall:1 knowledge:1 color:2 conversation:1 segmentation:3 actually:2 cbms:1 restarts:4 planar:1 wei:3 just:2 smola:1 hand:1 replacing:1 ei:1 nonlinear:1 indicated:1 lei:1 building:1 contain:1 true:4 hence:1 assigned:1 i2:2 deal:1 ll:1 noted:1 arrived:1 ay:1 jrl:1 image:1 recently:1 rotation:1 rl:1 empirically:1 volume:1 discussed:1 interpretation:1 extend:1 kluwer:1 interpret:1 ai:3 rd:1 meila:6 mathematics:1 similarly:1 zeroing:1 shawe:1 had:1 dj:5 moving:1 stable:2 similarity:2 surface:3 etc:1 closest:1 recent:1 apart:3 n00014:1 onr:1 success:2 a41:2 seen:1 minimum:3 r0:1 subblock:1 semi:1 ii:12 multiple:2 mix:1 academic:1 sphere:3 long:2 divided:1 equally:1 marina:1 laplacian:2 parenthesis:1 vision:4 kernel:6 normalization:2 whereas:2 want:2 fellowship:1 addressed:1 singular:2 appropriately:1 rest:1 regional:1 comment:1 tend:1 jordan:2 ideal:3 split:1 variety:1 xj:1 gave:2 idea:1 pca:2 eigengap:4 suffer:1 f:1 miiller:1 repeatedly:1 matlab:1 clear:1 eigenvectors:23 informally:1 ang:1 jrn:1 clusten:1 eory:1 sl:3 exist:2 nsf:1 s3:2 sjw:1 correctly:1 xii:1 discrete:2 alpert:1 four:2 threshold:1 christianini:1 graph:10 relaxation:1 padded:1 sum:6 run:4 family:1 reasonable:1 draw:1 dy:1 scaling:1 comparable:1 fl:3 guaranteed:1 annual:2 precisely:2 x2:1 n3:2 sake:1 nearby:1 attempting:1 vempala:1 according:2 yetta:1 poor:1 across:3 slightly:1 em:1 s1:6 intuitively:1 mutually:1 previously:3 discus:1 turn:2 mind:1 letting:2 flip:2 informal:1 adopted:1 available:2 tightest:1 unreasonable:1 apply:1 away:1 spectral:16 v2:1 alternative:2 yair:1 existence:1 original:2 top:1 clustering:24 cf:1 a4:3 unifying:1 giving:1 build:1 society:1 pled:1 malik:1 already:1 quantity:2 parametric:1 rt:1 diagonal:7 div:1 affinity:4 subspace:3 distance:3 link:1 thank:2 majority:1 outer:1 nx:1 seven:1 unstable:1 spanning:1 kannan:4 assuming:1 length:3 index:2 mini:1 ratio:2 hebrew:1 susceptible:1 relate:1 design:1 implementation:2 reliably:1 kahng:1 perform:1 disagree:1 finite:1 extended:1 rn:1 perturbation:6 varied:1 arbitrary:1 thm:1 required:1 specified:1 connection:1 optimized:1 usually:1 pattern:1 below:1 scott:1 rf:1 including:1 demanding:1 natural:2 improve:1 identifies:1 lk:1 harsh:1 sn:1 text:1 review:2 prior:1 relative:1 expect:2 proportional:1 foundation:2 fiji:1 degree:1 consistent:1 row:22 surprisingly:2 last:1 cohesiveness:1 supported:2 aij:5 side:1 understand:1 wide:1 fall:1 transition:2 contour:1 author:1 made:1 jump:1 refinement:1 simplified:1 far:2 sj:4 obtains:1 belongie:1 xi:3 search:1 iterative:1 why:1 promising:1 learn:1 reasonably:1 robust:1 parenthesized:1 longuet:1 cl:1 did:1 main:2 dense:1 s2:2 n2:3 repeated:2 eural:2 intel:1 fails:1 xl:1 lie:1 governed:1 perceptual:1 rk:2 theorem:2 transitioning:1 bad:2 british:1 maxi:2 symbol:2 dk:2 a3:3 grouping:1 exists:2 ih:1 texture:1 simply:1 infinitely:2 ordered:1 joined:1 subtlety:1 corresponds:1 chance:1 xix2:1 kli:1 kmeans:1 exposition:1 ajk:4 considerable:2 hard:2 experimentally:1 change:3 specifically:3 determined:1 except:1 principal:1 teng:1 experimental:1 e:1 formally:1 subblocks:1 latter:1 spielman:1 dept:1 instructive:1
1,199
2,093
Sequential noise compensation by sequential Monte Carlo method Kaisheng Yao and Satoshi Nakamura ATR Spoken Language Translation Research Laboratories 2-2-2, Hikaridai Seika-cho, Souraku-gun, Kyoto, 619-0288, Japan E-mail: {kaisheng.yao, satoshi.nakamura}@slt.atr.co.jp Abstract We present a sequential Monte Carlo method applied to additive noise compensation for robust speech recognition in time-varying noise. The method generates a set of samples according to the prior distribution given by clean speech models and noise prior evolved from previous estimation. An explicit model representing noise effects on speech features is used, so that an extended Kalman filter is constructed for each sample, generating the updated continuous state estimate as the estimation of the noise parameter, and prediction likelihood for weighting each sample. Minimum mean square error (MMSE) inference of the time-varying noise parameter is carried out over these samples by fusion the estimation of samples according to their weights. A residual resampling selection step and a Metropolis-Hastings smoothing step are used to improve calculation efficiency. Experiments were conducted on speech recognition in simulated non-stationary noises, where noise power changed artificially, and highly non-stationary Machinegun noise. In all the experiments carried out, we observed that the method can have significant recognition performance improvement, over that achieved by noise compensation with stationary noise assumption. 1 Introduction Speech recognition in noise has been considered to be essential for its real applications. There have been active research efforts in this area. Among many approaches, model-based approach assumes explicit models representing noise effects on speech features. In this approach, most researches are focused on stationary or slow-varying noise conditions. In this situation, environment noise parameters are often estimated before speech recognition from a small set of environment adaptation data. The estimated environment noise parameters are then used to compensate noise effects in the feature or model space for recognition of noisy speech. However, it is well-known that noise statistics may vary during recognition. In this situation, the noise parameters estimated prior to speech recognition of the utterances is possibly not relevant to the subsequent frames of input speech if environment changes. A number of techniques have been proposed to compensate time-varying noise effects. They can be categorized into two approaches. In the first approach, timevarying environment sources are modeled by Hidden Markov Models (HMM) or Gaussian mixtures that were trained by prior measurement of environments, so that noise compensation is a task of identification of the underlying state sequences of the noise HMMs, e.g., in [1], by maximum a posterior (MAP) decision. This approach requires making a model representing different conditions of environments (signal-to-noise ratio, types of noise, etc.), so that statistics at some states or mixtures obtained before speech recognition are close to the real testing environments. In the second approach, environment model parameters are assumed to be timevarying, so it is not only an inference problem but also related to environment statistics estimation during speech recognition. The parameters can be estimated by Maximum Likelihood estimation, e.g., sequential EM algorithm [2][3][4]. They can also be estimated by Bayesian methods. In the Bayesian methods, all relevant information on the set of environment parameters and speech parameters, which are denoted as ?(t) at frame t, is included in the posterior distribution given observation sequence Y (0 : t), i.e., p(?(t)|Y (0 : t)). Except for a few cases including linear Gaussian state space model (Kalman filter), it is formidable to evaluate the distribution updating analytically. Approximation techniques are required. For example, in [5], a Laplace transform is used to approximate the joint distribution of speech and noise parameters by vector Taylor series. The approximated joint distribution can give analytical formula for posterior distribution updating. We report an alternative approach for Bayesian estimation and compensation of noise effects on speech features. The method is based on sequential Monte Carlo method [6]. In the method, a set of samples is generated hierarchically from the prior distribution given by speech models. A state space model representing noise effects on speech features is used explicitly, and an extended Kalman filter (EKF) is constructed in each sample. The prediction likelihood of the EKF in each sample gives its weight for selection, smoothing, and inference of the time-varying noise parameter, so that noise compensation is carried out afterwards. Since noise parameter estimation, noise compensation and speech recognition are carried out frame-byframe, we denote this approach as sequential noise compensation. 2 Speech and noise model Our work is on speech features derived from Mel Frequency Cepstral Coefficients (MFCC). It is generated by transforming signal power into log-spectral domain, and finally, by discrete Cosine transform (DCT) to the cepstral domain. The following derivation of the algorithm is in log-spectral domain. Let t denote frame (time) index. In our work, speech and noise are respectively modeled by HMMs and a Gaussian mixture. For speech recognition in stationary additive noise, the following formula [4] has been shown to be effective in compensating noise effects. For Gaussian mixture kt at state st , the Log-Add method transforms the mean vector ?lst kt of the Gaussian mixture by, ? ?lst kt = ?lst kt + log(1 + exp(?ln ? ?lst kt )) (1) where ?ln is the mean vector in the noise model. st ? {1, ? ? ? , S}, kt ? {1, ? ? ? , M }. S and M each denote the number of states in speech models and the number of mixtures at each state. Superscript l indicates that parameters are in the logspectral domain. After the transformation, the mean vector ? ? lst kt is further transformed by DCT, and then plugged into speech models for recognition of noisy speech. In case of time-varying noise, the ?ln should be a function of time, i.e., ?ln (t). Accordingly, the compensated mean is ? ? lst kt (t). s0 st ?1 st sT k0 kt ?1 kt kT ? sl 0 k 0 (0) ? sl t ?1kt ?1 (t ? 1) Y l ( 0) ? nl (0) ? sl t k t (t ) Y l (t ? 1) ? nl (t ? 1) ? sl T kT (T ) Y l (t ) ? nl (t ) Y l (T ) ? nl (T ) Figure 1: The graphical model representation of the dependences of the speech and noise model parameters. st and kt each denote the state and Gaussian mixture at frame t in speech models. ?lst kt (t) and ?ln (t) each denote the speech and noise parameter. Y l (t) is the noisy speech observation. The following analysis can be viewed in Figure 1. In Gaussian mixture kt at state st of speech model, speech parameter ?lst kt (t) is assumed to be distributed in Gaussian with mean ?lst kt and variance ?lst kt . On the other hand, since the environment parameter is assumed to be time varying, the evolution of the environment mean vector can be modeled by a random walk function, i.e., ?ln (t) = ?ln (t ? 1) + v(t) (2) where v(t) is the environment driving noise in Gaussian distribution with zero mean and variance V . Then, we have, p(st , kt , ?lst kt (t), ?ln (t)|st?1 , kt?1 , ?lst?1 kt?1 (t ? 1), ?ln (t ? 1)) = ast?1 st pst kt N (?lst kt (t); ?lst kt , ?lst kt )N (?ln (t); ?ln (t ? 1), V ) (3) where ast?1 st is the state transition probability from st?1 to st , and pst kt is the mixture weight. The above formula gives the prior distribution of the set of speech and noise model parameter ?(t) = {st , kt , ?lst kt (t), ?ln (t)}. Furthermore, given observation Y l (t), assume that the transformation by (1) has modeling and measurement uncertainty in Gaussian distribution, i.e., Y l (t) = ?lst kt (t) + log (1 + exp (?ln (t) ? ?lst kt (t))) + wst kt (t) (4) where wst kt (t) is Gaussian with zero mean and variance ?lst kt , i.e., N (?; 0, ?lst kt ). Thus, the likelihood of observation Y l (t) at state st and mixture kt is p(Y l (t)|?(t)) = N (Y l (t); ?lst kt (t) + log (1 + exp (?ln (t) ? ?lst kt (t))), ?lst kt ) (5) Refereeing to (3) and (5), the posterior distribution of ?(t) given Y l (t) is p(st , kt , ?lst kt (t), ?ln (t)|Y l (t)) ? p(Y l (t)|?(t))ast?1 st pst kt N (?lst kt (t); ?lst kt , ?lst kt )N (?ln (t); ?ln (t ? 1), V ) (6) The time-varying noise parameter is estimated by MMSE, given as, Z XZ l l p(?(t)|Y l (0 : t))d?lst kt (t)d?ln (t) ?n (t) ? ?n (t) = ?ln (t) st ,kt ?ls t kt (7) (t) However, it is difficult to obtain the posterior distribution p(?(t)|Y l (0 : t)) analytically, since p(?lst kt (t), ?ln (t)|Y l (t)) is non-Gaussian in ?lst kt (t) and ?ln (t) due to the non-linearity in (4). It is thus difficult, if possible, to assign conjugate prior of ?ln (t) to the likelihood function p(Y l (t)|?(t)). Another difficulty is that the speech state and mixture sequence is hidden in (7). We thus rely on the solution by computational Bayesian approach [6]. 3 Time-varying noise parameter estimation by sequential Monte Carlo method We apply the sequential Monte Carlo method [6] for posterior distribution updating. At each frame t, a proposal importance distribution is sampled whose target is the posterior distribution in (7), and it is implemented by sampling from lower distributions in hierarchy. The method goes through the sampling, selection, and smoothing steps frame-by-frame. MMSE inference of the time-varying noise parameter is a by-product of the steps, carried out after the smoothing step. In the sampling step, the prior distribution given by speech models is set to the proposal importance distribution, i.e., q(?(t)|?(t ? 1)) = ast?1 st pst kt N (?lst kt (t); ?lst kt , ?lst kt ). The samples are then generated by sampling hierarchically of the prior distribution described as follows: set i = 1 and perform the following steps: (i) 1. sample st ? as(i) s t?1 t 2. sample (i) kt 3. sample ? ? ps(i) kt l(i) (i) (i) st kt t (t) ? N (; ?l (i) (i) st kt , ?l (i) (i) st kt ), and set i = i + 1 4. repeat step 1 to 3 until i = N where superscript (i) denotes the index of samples and N denotes the number of samples. Each sample represents certain speech and noise parameter, which is (i) (i) l(i) l(i) denoted as ?(i) (t) = (st , kt , ? (i) (i) (t), ?n (t)). The weight of each sample is st kt Qt p(?(? )(i) |Y l (? )) given by ? =1 q(?(? . Refereeing to (6), the weight is calculated by )(i) |?(? ?1)(i) ) l(i) ?(i) ? (i) (t) = p(Y l (t)|?(i) (t))N (?l(i) n (t); ?n (t ? 1), V )? (t ? 1) (8) where ??(i) (t ? 1) is the sample weight from previous frame. The remaining part in the right side of above equation, in fact, represents the prediction likelihood of the state space model given by (2) and (4) for each sample (i). This likelihood can be obtained analytically since after linearization of (4) with respect to ?ln (t) at l(i) ?n (t ? 1), an extended Kalman filter (EKF) can be obtained, where the prediction likelihood of the EKF gives the weight, and the updated continuous state of EKF l(i) gives ?n (t). In practice, after the above sampling step, the weights of all but several samples may become insignificant. Given the fixed number of samples, this will results in degeneracy of the estimation, where not only some computational resources are wasted, but also estimation might be biased because of losing detailed information on some parts important to the parameter estimation. A selection step by residual resampling [6] is adopted after the sampling step. The method avoids the degeneracy by discarding those samples with insignificant weights, and in order to keep the number of the samples constant, samples with significant weights are duplicated. Accordingly, the weights after the selection step are also proportionally redistributed. Denote the ? ? (i) (t); i = 1 ? ? ? N } with weights set of samples after the selection step as ?(t) = {? ? = {??(i) (t); i = 1 ? ? ? N }. ?(t) After the selection step at frame t, these N samples are distributed approximately according to the posterior distribution in (7). However, the discrete nature of the approximation can lead to a skewed importance weights distribution, where ? the extreme case is all the samples have the same ?(t) estimated. A MetropolisHastings smoothing [7] step is introduced in each sample where the step involves ? (i) (t) according to the proposal sampling a candidate ??(i) (t) given the current ? ? (i) ? (t)). The Markov chain then moves towards importance distribution q(? (t)|? ?(i) l ? (i) ?(i) |Y (t))q(? |? ) ??(i) (t) with acceptance possibility as min{1, p(? ? (i) |Y l (t))q(??(i) |? ? (i) ) }, otherwise it p(? ? (i) . To simplify calculation, we assume that the importance distriburemains at ? ? ? (i) (t)) is symmetric, and after some mathematical manipulation, it tion q(? (t)|? ?(i) (t) is shown that the acceptance possibility is given by min{1, ???(i) (t) }. Denote the (i) ? ? ? obtained samples as ?(t) = {? (t); i = 1 ? ? ? N } with weights ?(t) = {??(i) (t); i = 1 ? ? ? N }. Noise parameter ?ln (t) is estimated via MMSE over the samples, i.e., ? ?ln (t) = N X i=1 l(i) ??(i) (t) ?l(i) PN ?(j) ? n (t) (t) j=1 ? where ? ?n (t) is the updated continuous state of the EKF in the sample after the smoothing step. Once the estimate ? ? ln (t) has been obtained, it is plugged into (1) to do non-linear transformation of clean speech models. 4 4.1 Experimental results Experimental setup Experiments were performed on the TI-Digits database down-sampled to 16kHz. Five hundred clean speech utterances from 15 speakers and 111 utterances unseen in the training set were used for training and testing, respectively. Digits and silence were respectively modeled by 10-state and 3-state whole word HMMs with 4 diagonal Gaussian mixtures in each state. The window size was 25.0ms with a 10.0ms shift. Twenty-six filter banks were used in the binning stage. The features were MFCC + ? MFCC. The baseline system had a 98.7% Word Accuracy under clean conditions. We compared three systems. The first was the baseline trained on clean speech without noise compensation, and the second was the system with noise compensation by (1) assuming stationary noise [4]. They were each denoted as Baseline and Stationary Compensation. The sequential method was un-supervised, i.e., without training transcript, and it was denoted according to the number of samples and variance of the environment driving noise V . Four seconds of contaminating noise was used in each experiment to obtain noise mean vector ?ln in (1) for Stationary Compensation. It was also for initialization of ?ln (0) in the sequential method. The initial l(i) ?n (0) for each sample was sampled from N (?ln (0), 0.01) + N (?ln (0) + ?(0), 10.0), where ?(0) was flat distribution in [?1.0, 9.0]. 4.2 Speech recognition in simulated non-stationary noise White noise signal was multiplied by a Chirp signal and a rectangular signal, so that the noise power of the contaminating White noise changed continuously, denoted as experiment A, and dramatically, denoted as experiment B. As a result, signalto-noise ratio (SNR) of the contaminating noise ranged from 0dB to 20.4dB. We plotted the noise power in 12th filter bank versus frames in Figure 2, together with the estimated noise power by the sequential method with number of samples set to 120 and environment driving noise variance set to 0.0001. As a comparison, we also plotted the noise power and its estimate by the method with the same number of samples but larger driving noise variance to 0.001. By Figure 2 and Figure 3, we have the following observations. First, the method can track the evolution of the noise power. Second, the larger driving noise variance V will make faster convergence but larger estimation error of the method. In terms of recognition performance, Table 1 shows that the method can effectively improve system robustness to the time-varying noise. For example, with 60 samples, and the environment driving noise variance V set to 0.001, the method can improve word accuracy from 75.30% achieved by ?Stationary Compensation?, to 94.28% in experiment A. The table also shows that, the word accuracies can be improved by increasing number of samples. For example, given environment driving noise variance V set to 0.0001, increasing number of samples from 60 to 120, can improve word accuracy from 77.11% to 85.84% in experiment B. Table 1: Word Accuracy (in %) in simulated non-stationary noises, achieved by the sequential Monte Carlo method in comparison with baseline without noise compensation, denoted as Baseline, and noise compensation assuming stationary noise, denoted as Stationary Compensation. Experiment Baseline Stationary Compensation A B 48.19 53.01 75.30 78.01 4.3 # samples = 60 V 0.001 0.0001 94.28 93.98 82.23 77.11 # samples = 120 V 0.001 0.0001 94.28 94.58 85.84 85.84 Speech recognition in real noise In this experiment, speech signals were contaminated by highly non-stationary Machinegun noise in different SNRs. The number of samples was set to 120, and the environment driving noise variance V was set to 0.0001. Recognition performances are shown in Table 2, together with ?Baseline? and ?Stationary Compensation?. Figure 2: Estimation of the time-varying parameter ?ln (t) by the sequential Monte Carlo method at 12th filter bank in experiment A. Number of samples is 120. Environment driving noise variance is 0.0001. Solid curve is the true noise power. Dash-dotted curve is the estimated noise power. It is observed that, in all SNR conditions, the method can further improve system performance, compared to that obtained by ?Stationary Compensation?, over ?Baseline?. For example, in 8.86dB SNR, the method can improve word accuracy from 75.60% by ?Stationary Compensation? to 83.13%. As a whole, the method can have a relative 39.9% word error rate reduction compared to ?Stationary Compensation?. Table 2: Word Accuracy (in %) in Machinegun noise, achieved by the sequential Monte Carlo method in comparison with baseline without noise compensation, denoted as Baseline, and noise compensation assuming stationary noise, denoted as Stationary Compensation. SNR (dB) 28.86 14.88 8.86 1.63 5 Baseline 90.36 64.46 56.02 50.0 Stationary Compensation 92.77 76.81 75.60 68.98 #samples = 120, V = 0.0001 97.59 88.25 83.13 72.89 Summary We have presented a sequential Monte Carlo method for Bayesian estimation of time-varying noise parameter, which is for sequential noise compensation applied to robust speech recognition. The method uses samples to approximate the posterior distribution of the additive noise and speech parameters given observation sequence. Figure 3: Estimation of the time-varying parameter ?ln (t) by the sequential Monte Carlo method at 12th filter bank in experiment A. Number of samples is 120. Environment driving noise variance is 0.001. Solid curve is the true noise power. Dash-dotted curve is the estimated noise power. Once the noise parameter has been inferred, it is plugged into a non-linear transformation of clean speech models. Experiments conducted on digits recognition in simulated non-stationary noises and real noises have shown that the method is very effective to improve system robustness to time-varying additive noise. References [1] A. Varga and R.K. Moore, ?Hidden markov model decomposition of speech and noise,? in ICASSP, 1990, pp. 845?848. [2] N.S. Kim, ?Nonstationary environment compensation based on sequential estimation,? IEEE Signal Processing Letters, vol. 5, no. 3, March 1998. [3] K. Yao, K. K. Paliwal, and S. Nakamura, ?Sequential noise compensation by a sequential kullback proximal algorithm,? in EUROSPEECH, 2001, pp. 1139?1142, extended paper submitted for publication. [4] K. Yao, B. E. Shi, S. Nakamura, and Z. Cao, ?Residual noise compensation by a sequential em algorithm for robust speech recognition in nonstationary noise,? in ICSLP, 2000, vol. 1, pp. 770?773. [5] B. Frey, L. Deng, A. Acero, and T. Kristjansson, ?Algonquin: Iterating laplace?s method to remove multiple types of acoustic distortion for robust speech recognition,? in EUROSPEECH, 2001, pp. 901?904. [6] J. S. Liu and R. Chen, ?Sequential monte carlo methods for dynamic systems,? J. Am. Stat. Assoc, vol. 93, pp. 1032?1044, 1998. [7] W. K. Hastings, ?Monte carlo sampling methods using markov chains and their applications,? Biometrika, vol. 57, pp. 97?109, 1970.
2093 |@word kristjansson:1 decomposition:1 solid:2 reduction:1 initial:1 liu:1 series:1 mmse:4 current:1 dct:2 subsequent:1 additive:4 remove:1 resampling:2 stationary:23 accordingly:2 five:1 mathematical:1 constructed:2 become:1 seika:1 xz:1 compensating:1 window:1 increasing:2 underlying:1 linearity:1 formidable:1 evolved:1 spoken:1 transformation:4 ti:1 biometrika:1 assoc:1 before:2 frey:1 approximately:1 might:1 initialization:1 chirp:1 co:1 hmms:3 testing:2 practice:1 digit:3 area:1 word:9 close:1 selection:7 acero:1 ast:4 map:1 compensated:1 shi:1 go:1 l:1 focused:1 rectangular:1 laplace:2 updated:3 target:1 hierarchy:1 losing:1 us:1 recognition:21 approximated:1 updating:3 database:1 binning:1 observed:2 environment:22 transforming:1 dynamic:1 trained:2 efficiency:1 icassp:1 joint:2 k0:1 lst:33 derivation:1 snrs:1 effective:2 monte:12 whose:1 larger:3 distortion:1 otherwise:1 statistic:3 unseen:1 transform:2 noisy:3 superscript:2 sequence:4 analytical:1 redistributed:1 product:1 adaptation:1 relevant:2 cao:1 convergence:1 p:1 generating:1 stat:1 qt:1 transcript:1 implemented:1 involves:1 filter:8 assign:1 icslp:1 considered:1 exp:3 driving:10 vary:1 estimation:16 gaussian:13 ekf:6 pn:1 varying:15 timevarying:2 publication:1 derived:1 improvement:1 likelihood:8 indicates:1 baseline:11 kim:1 am:1 inference:4 hidden:3 transformed:1 among:1 denoted:10 smoothing:6 once:2 sampling:8 represents:2 report:1 contaminated:1 simplify:1 few:1 acceptance:2 highly:2 possibility:2 mixture:12 extreme:1 nl:4 chain:2 algonquin:1 kt:62 wst:2 taylor:1 plugged:3 walk:1 plotted:2 modeling:1 snr:4 hundred:1 conducted:2 eurospeech:2 proximal:1 cho:1 st:25 together:2 continuously:1 yao:4 possibly:1 japan:1 coefficient:1 explicitly:1 tion:1 performed:1 square:1 accuracy:7 variance:12 satoshi:2 identification:1 bayesian:5 carlo:12 mfcc:3 submitted:1 frequency:1 pp:6 degeneracy:2 sampled:3 pst:4 duplicated:1 supervised:1 improved:1 furthermore:1 stage:1 until:1 hand:1 hastings:2 effect:7 ranged:1 true:2 evolution:2 analytically:3 symmetric:1 laboratory:1 moore:1 white:2 during:2 skewed:1 speaker:1 mel:1 cosine:1 m:2 khz:1 jp:1 significant:2 measurement:2 refereeing:2 language:1 had:1 etc:1 add:1 posterior:9 contaminating:3 manipulation:1 paliwal:1 certain:1 minimum:1 deng:1 signal:7 afterwards:1 multiple:1 kyoto:1 faster:1 calculation:2 compensate:2 prediction:4 achieved:4 proposal:3 source:1 biased:1 db:4 nonstationary:2 shift:1 six:1 effort:1 speech:47 dramatically:1 varga:1 detailed:1 proportionally:1 iterating:1 transforms:1 sl:4 dotted:2 estimated:11 track:1 discrete:2 vol:4 four:1 clean:6 wasted:1 letter:1 uncertainty:1 decision:1 dash:2 flat:1 generates:1 min:2 according:5 march:1 conjugate:1 em:2 metropolis:1 making:1 ln:32 equation:1 resource:1 adopted:1 multiplied:1 apply:1 spectral:2 alternative:1 robustness:2 assumes:1 denotes:2 remaining:1 graphical:1 hikaridai:1 move:1 kaisheng:2 dependence:1 diagonal:1 atr:2 simulated:4 hmm:1 gun:1 mail:1 assuming:3 kalman:4 modeled:4 index:2 ratio:2 difficult:2 setup:1 twenty:1 perform:1 observation:6 markov:4 compensation:29 situation:2 extended:4 frame:11 inferred:1 introduced:1 required:1 acoustic:1 including:1 power:11 metropolishastings:1 difficulty:1 nakamura:4 rely:1 residual:3 representing:4 improve:7 carried:5 utterance:3 prior:9 relative:1 versus:1 s0:1 bank:4 translation:1 changed:2 summary:1 repeat:1 silence:1 side:1 cepstral:2 slt:1 distributed:2 curve:4 calculated:1 transition:1 avoids:1 approximate:2 kullback:1 keep:1 active:1 assumed:3 continuous:3 un:1 table:5 nature:1 robust:4 artificially:1 domain:4 hierarchically:2 whole:2 noise:98 categorized:1 slow:1 explicit:2 candidate:1 weighting:1 formula:3 down:1 discarding:1 insignificant:2 fusion:1 essential:1 sequential:22 effectively:1 importance:5 linearization:1 chen:1 signalto:1 viewed:1 towards:1 change:1 included:1 except:1 experimental:2 evaluate:1