A Cortex-inspired Associative Memory with O(1) Time Complexity Learning, Recall, and Recognition of Sequences

Size: px
Start display at page:

Download "A Cortex-inspired Associative Memory with O(1) Time Complexity Learning, Recall, and Recognition of Sequences"

Transcription

1 A Corex-inspired Associaive Memory wih O(1) Time Complexiy Learning, Recall, and Recogniion of Sequences Gerard J. Rinkus Volen Cener for Complex Sysems Brandeis Universiy Absrac A corex-inspired associaive memory model possessing O(1) ime complexiy for boh sorage (learning) and rerieval (recall, recogniion) of sequences is described. I learns sequences, specifically, binary spaioemporal paerns, wih single rials. Resuls are given demonsraing O(1) rerieval of: a) a sequence s remaining iems when promped wih is iniial iem, i.e., episodic recall; and b) he mos similar sored sequence when presened wih a novel sequence, i.e., recogniion/caegorizaion. The hidden (represenaion) layer, L2, is organized ino winner-ake-all compeiive modules (CMs) hypohesized o be analogous o corical minicolumns. Represenaions consis of one acive uni per CM. The hear of he model is a maching algorihm ha, given he curren inpu momen, σ, i.e., he curren inpu iem in he conex of he sequence hus far: a) finds he sored represenaion, σ*, of ha previously experienced momen, σ*, ou of all previously experienced momens, which is spaioemporally mos similar o σ; and b) reurns a normalized measure, G, of ha similariy. When in recall or recogniion mode, he model simply reacivaes σ* since i is he mos likely hypohesis given he model s hisory and he curren inpu. When in learning mode, he model injecs an amoun of noise, inversely proporional G, ino he process of choosing he cells o represen σ. This yields he propery ha he size of he inersecion beween represenaions is an increasing funcion of he spaioemporal similariy of he momens ha hey represen. Thus, he higher-order saisics (spaioemporal similariy srucure) of he se of learned sequences is refleced direcly in he paerns of inersecions over he se of represenaions. This propery, in conjuncion wih he use of he binary sparse represenaion, makes he O(1) recall and recogniion (i.e., inference) possible. 1

2 1. Inroducion Human capabiliy sill far exceeds ha of machines on a vas range of informaion processing asks. Parsing speech in noisy environmens, undersanding he meaning of ha speech a many levels from he lieral up hrough recognizing sarcasm or irony, reading an opponen s inenion, and hus being able o predic his or his eam s nex or ensuing movemens, ec. These are all spaioemporal (sequenial) asks ha humans wih sufficien experience rouinely perform boh accuraely and in real ime. An underlying commonaliy beween asks like hese is ha hey all require rapid inference in exponenially large hypohesis spaces. For example, he space of possible movemen paerns over he nex five seconds of he se of opposing soccer eam players, is exponenially large. Neverheless, an experienced player can ofen correcly predic he paern wih sufficien accuracy o be successful in he momen. The relaively slow speed of neural operaions implies ha somehow he brain drasically pares down he huge hypohesis space so ha very few hypoheses are explicily considered a any given momen. A base, he abiliy o achieve such a paring down is a funcion of how he hypoheses are represened in memory. For example, he log 2 N ime complexiy of binary search depends on he fac ha he se of N iems being searched over is arranged by numerical similariy, e.g.., 8 is more similar o 9 han o 2. More generally, one migh say ha search and rerieval is abeed o he exen ha he similariy srucure of he hypohesis space is refleced in he physical represenaion (sorage) of he hypoheses. Of course, ensuring ha such a propery i.e., similar inpus map o similar represenaions (SISR) holds as successive inpus are presened o he sysem generally incurs increased sorage-ime complexiy. However, he model described herein implemens his propery in such a way ha i achieves he heoreically opimal ime complexiy, i.e., O(1), for boh sorage and rerieval. This model, originally developed in Rinkus (1995, 1996), employs sparse disribued represenaions o do unsupervised learning, recall and recogniion of binary spaioemporal paerns, or sequences, and possesses his immediae search/rerieval propery. More formally, he model has O(1) search ime complexiy, meaning ha he number of compuaional seps necessary o find he mos similar sored sequence (o some query sequence) remains consan no maer how many sequences have been sored, and no maer how long individual sequences are. This holds 2

3 wheher he model is asked o recall he remainder of a sequence given a promp (episodic recall) or recognize a novel sequence (by acivaing he memory race of he mos similar previously learned sequence). Moreover, he model does single-rial learning of is inpu sequences and, in fac, has O(1) sorage ime complexiy as well. The model, named TEMECOR (TEmporal MEmory using COmbinaorial Represenaions), achieves his performance primarily hrough wo feaures: a) a paricular, sparse disribued, k-wta, represenaion, modeled afer he minicolumn srucure of neocorex; and b) an algorihm which compues, a each momen during learning, he maximal similariy, G, beween he currenly presening sequence and all of is sored sequences (and all of heir prefixes) and hen adds an amoun of noise, inversely proporional o G, ino he choice of unis ha will become acive a ha momen. Togeher, hese feaures yield he key propery ha he degree of overlap beween wo sored sequence represenaions (memory races) is an increasing funcion of he spaioemporal similariy of he sequences ha hey represen, i.e., he SISR propery. Thus, boh he memory races of he individual sequences and he higher-order saisics (similariy srucure) of he se of sequences reside in a single monolihic srucure (se of weighs). The specific claim of his paper is ha he proposed model has O(1) compuaional ime complexiy for boh sorage (learning) and rerieval (recall, recogniion) of sequences. This follows immediaely from he fac ha none of he model s equaions (Eqs. 1-11) have any dependence on he number sequences sored. While he compuaional space complexiy, i.e., sorage capaciy, of an associaive memory model is also of paramoun imporance, his paper does no presen a definiive resul for he model s space complexiy. However, he appendix gives wo indirec argumens suggesing ha he model s space complexiy will allow scaling o biologically relevan problem sizes. A comprehensive examinaion of he model s space complexiy will be presened in a fuure paper. 2. Background The use of disribued represenaions for caegories (hidden variables, causes, facors) has been cenral o boh he mulilayer percepron (MLP), e.g., Backpropagaion, and associaive memory approaches o compuaional inelligence. However, his is no rue of he family of probabilisic models, generally referred o as graphical models, which has become he dominan paradigm in he field. This very large family includes Bayesian Belief Neworks (Pearl, 1988), hidden Markov models (HMMs), 3

4 mixure models, principal componens analysis, amongs ohers (Roweis & Ghahramani, 1999). All of hese models were originally described wih he assumpion of a localis (singleon) represenaion of caegories. Many of hem, noably HMMs, which is by far, he predominan approach for speech recogniion (and perhaps for he domain of emporal/sequenial paern recogniion in general), have reained his localis assumpion over almos heir enire developmenal courses. More recenly, here has been a movemen owards disribued represenaions for caegories in graphical models. This movemen is ofen aribued o he observaion, by Williams & Hinon (1991), of a fundamenal inefficiency of localis represenaions. They observed ha, in order o represen only N bis of informaion abou he hisory of a ime series, he single mulinomial sae variable of a sandard, localis HMM requires 2 N possible values, i.e., an HMM wih 2 N nodes. In conras, a disribued represenaion of sae wih only N binary variables (an HMM wih 2N nodes) can also sore N bis of hisory informaion. A number of models using disribued represenaions of sae have since been proposed (Ghahramani & Jordan, 1997; Brown & Hinon, 2001; Jacobs e al., 2002; Blizer e al., 2005). The benefi of disribued represenaions can be viewed as increased capaciy for soring he higher-order saisics presen in a sequence (or in any daa se, more generally). This poin can be made more concreely in erms of sparse disribued represenaions. In a sparse disribued represenaion, in which any given represenee is represened by a small subse of represenaional unis (e.g., 100 ou of 1000) being acive, he higher-order saisics, i.e., similariy srucure, of he se of represenees, can be represened direcly by he paerns of overlaps (inersecions) over heir represenaions. The more similar wo represenees are, he more unis hey have in common. Figure 1 illusraes his concep. A lef is a hypoheical se of observaions (A, B, D) in some feaure space and a hypoheical underlying caegory srucure (C1, C2) consisen wih he disances beween he observaions in he feaure space. A lower righ is one possible sparse disribued represenaion: each observaion is represened by 5 ou of 12 of he unis (which is sparse enough o make he poin). The verical doed lines on he righ show how he paern of overlaps of he hree represenaions can direcly represen he higher-order saisics (a nesed caegorical srucure) of he observaions: he more similar observaions, A and B, have hree unis in common, which consiues a represenaion of subcaegory, C1, while he full se of observaions have only wo unis in common, which consiues a represenaion of he super-caegory, C2. 4

5 Hidden, hough acual, caegorical srucure Paern of overlaps, which implicily represens he similariy srucure of he inpu space. animals C2 C2 fish C1 C1 A A B D rou shark bear B D Observaions Represenaions of he individual observaions Figure 1: (Lef) A hypoheical se of hree observed paerns (A, B, D) from some inpu space and a hypoheical underlying caegory srucure. A and B are more similar han A and D or B and D. (Lower righ) A sparse disribued represenaion of he individual observaions, A, B, and D. (Upper righ) Differen subses of unis, C1 and C2, corresponding o he pair-wise and riple-wise overlaps respecively, consiue explici represenaions of higher-order similariy (i.e., caegory) srucure. Figure 1 is inended o show only ha i is possible o represen he similariy srucure of a se of individual observaion by he paern of overlaps over heir represenaions. The informaion presen in he paern of overlaps in he box a upper righ is fully analogous o a dendrogram (hierarchical clusering analysis) ypically used o demonsrae he caegorizaion behavior of learning models ha use fully disribued represenaions (Elman, 1990). In he case of sparse disribued represenaions, as in Figure 1, he appropriae measure of disance beween represenaions is Hamming disance. The paern of overlaps in he upper righ box is isomorphic o a dendrogram in which A and B would be clusered as C1 and hen C1 and D would be clusered as C2, which would look exacly like he ree srucure on he lef, as desired. The evidence, given in Secion 4, ha he proposed model acually uses his implici caegory informaion is ha i correcly classifies novel observaion (sequences) even hough here are no explici and separae represenaions of caegory informaion in he model. 5

6 The represenaion of similariy srucure by paerns of overlap consiues nohing less han an addiional represenaional dimension, one which is unavailable a eiher exreme of he represenaional coninuum, i.e., in eiher localis or densely (fully) disribued represenaions. Under a localis scheme here are no code overlaps; hus, he only way o represen he higher-order saisical srucure over he represenees is in he weighs beween he represenaions. Under a dense scheme, where formally, all unis paricipae in all represenaions, all code overlaps are complee, e.g., a hidden layer wih M real-valued unis in a Backpropagaion model. Here, he only available represenaion of similariy is a scalar (e.g., Euclidean) disance beween he represenaions. 1 The possibiliy of represening similariy by overlaps has been generally underappreciaed hus far. However, i is a he core of he spaial model of Rachkovskij & Kussel (2000, 2001) and of he spaioemporal model proposed herein. TEMECOR has no been developed wihin he framework of graphical models. In paricular, alhough i deals wih spaioemporal paerns, i has no been developed as an exension of any varian of HMMs, nor of he more general class of Dynamic Belief Nes (DBNs). Is learning algorihm will, hus, no be cas as a version of he expecaion-maximizaion (EM) algorihm (Dempser & Laird, 1977). Insead, he learning algorihm, which uses only single rials, will be described as compuing a spaioemporal similariy meric and using i o guide he choice of represenaion so ha he paern of overlaps over he final se of represenaions reflecs he higher-order spaioemporal saisics of he inpu paern se. As will be shown, his propery leads direcly o a simple rerieval (inference) algorihm, ha finds he mos similar sored sequence o a query sequence (i.e., he maximum likelihood hypohesis) in a single compuaion, i.e., a compuaion which does no depend on how many sequences have been sored.. TEMECOR has also no been exended from he various MLP varians ha have been applied o emporal/sequenial paerns, e.g., Waibel (1989), Jordan (1986), Elman (1990), Williams & Zipser (1989). The MLP-based approach has been successfully applied o many caegorizaion problems. 1 This issue is cenral o composiionaliy. Classical AI approaches are localis. Composiionaliy is implicily presen owing o he clear mapping beween represenees, boh wholes and pars, and heir represenaions. For connecionis approaches ha employ dense disribuions, here is no such clear mapping available, hence he classicis s (Fodor & Pylyshyn, 1988) claim ha connecionis sysems lack wha Van Gelder (1990) erms synacic composiionaliy. However, as discussed here, sparse connecionis models can possess synacic composiionaliy. 6

7 However, such models have: a) no been demonsraed o simulaneously explain boh episodic memory (recall of sequences) and semanic memory (caegorizaion of novel sequences); b) require numerous raining rials of each exemplar; and c) suffer from caasrophic forgeing in which nearly all informaion abou a previously learned daa se can be los when he model learns a new daa se (McCloskey & Cohen, 1989; French, 1991; Cleeremans, 1993). Moreover, as discussed by Cleeremans, such Backpropagaion-based models have difficulies ha are paricularly exacerbaed in he conex of learning complex sequences, which are sequences in which iems can recur muliple imes and in varying conexs. TEMECOR is closer o a number of sparse disribued associaive memory models (e.g., Willshaw e al., 1969; Palm, 1980; Lynch, 1986; Kanerva, 1988; Moll & Miikkulainen, 1996), alhough i differs from hese in many ways, including is focus on spaioemporal paerns and is novel use of noise o effec he crucial SISR (similar inpus map o similar represenaions) propery. I also has similariies o he Synfire Chains model (Abeles, 1991) in ha i maps inpus, specifically sequences, ino spaioemporal memory races, alhough he focus is no on he fine-scale iming of spikes. Several auhors, e.g., Dayan & Zemel (1995), have discussed he opposiional naures of he coding mechanisms of cooperaion and compeiion. O Reilly (1998) explains ha any use of compeiion prevens he cooperaiviy and combinaorialiy of rue disribued represenaions, and he need o preserve independence among he unis in a [disribued represenaion] prevens he inroducion of any rue acivaion-based compeiion. He adds ha he developmen of sparse disribued represenaions, which necessarily inegrae hese wo opposing mechanisms, has been challenging largely because hey are difficul o analyze mahemaically. Relaedly, a number of auhors have discussed he benefis of and need for sparse disribued represenaions (Olshausen & Field, 1996; Hinon & Ghahramani, 1997). TEMECOR may provide a hopeful alernaive in his regard because is paricular implemenaion of he sparse disribued paradigm cleanly separaes he cooperaive and compeiive aspecs. As will be seen shorly, compeiion is isolaed wihin he model s winner-ake-all compeiive modules (CMs), and he cooperaion is presen in ha inernal represenaions (codes) are ses of coacive unis, one per CM. 3. Model Descripion Figure 2 shows he ypes of sequences processed by he model. In his example, he inpu surface consiss of 16 absrac binary feaures arranged 7

8 in a verical array; array posiion does no reflec any spaial relaion in he physical inpu domain. Figure 2a shows an insance of a compleely uncorrelaed or whie noise sequence in which a randomly chosen subse of binary feaures is chosen o be acive a each momen.. The inenion of using inpu ses consising of such uncorrelaed sequences is o approximae he condiions of episodic memory wherein essenially arbirary combinaions of feaures are remembered essenially permanenly despie occurring only once. The model also handles correlaed, or more specifically, complex, sequence ses (Figure 2b) in which whole inpu iems can recur muliple imes and in differen emporal conexs, as is necessary for modeling language-like domains. a) Uncorrelaed b) Correlaed (Complex) 3 10 L O T T O Figure 2: a) A random binary spaioemporal paern (sequence) having 10 ime seps, or momens. The inpu surface has 16 absrac binary feaures, a randomly chosen subse of which is acive a each momen. b) A complex sequence in which he same iem (subse of feaures), e.g., an English leer, can recur in differen conexs, as is appropriae for represening linguisic domains. As shown in Figure 3, he wo-layer model 2 uses sparse disribued binary represenaions, or codes, in is inernal layer, L2. In his example, L2 is divided up ino Q = 4 winner-ake-all compeiive modules (CMs) and every code consiss of one acive uni in each CM (black L2 unis). Hence, he layer insaniaes a paricular ype of k-wta archiecure. Here, L1 is arranged as a 2D grid of binary absrac feaures; he model is agnosic o modaliy and o any inheren dimensionaliy of he inpu space. Figure 3 shows he sae of a small insance of he model on all four ime seps 2 This model has already been generalized o a hierarchical framework allowing an arbirary number of layers (Rinkus & Lisman, 2005) and will be described in a separae manuscrip. 8

9 during he learning of he sequence, [ABDB]. I illusraes several aspecs of he model s basic operaion, as well as some nomenclaure. Noe ha only a small subse of he connecions (weighs) are shown in his figure. Throughou his paper, we will assume complee iner-level conneciviy: all L1 unis connec wih all L2 unis in boh he boom-up (BU) and op-down (TD) direcion. We also assume nearly complee inra-level (horizonal, or H) conneciviy (wihin L2 only): all L2 unis conac all oher L2 unis excep hose in is own CM. Code: A AB ABD ABDB L2 L1 Time: Inpu: Momen: 1 A [A] 2 B [AB] Figure 3: A depicion of a small insance of he model on all four ime seps, or momens, while learning he sequence, [ABDB]. Only a represenaive subse of he weigh increases on each ime sep are shown (black lines). See ex for deailed explanaion of he figure. Figure 3 shows ha during learning, a new L2 code generally becomes acive for each successive ime sep, or momen, of a sequence. A momen is defined as a paricular spaial inpu (iem) in he conex of he paricular sequence of iems leading up o i (i.e., a paricular prefix or hisory). Thus, he sequence [ABDB] consiss of four unique momens, [A], [AB], [ABD], and [ABDB], bu only hree unique inpu iems. The symbol, σ, is used o denoe a momen. The symbol, ϒ, is used o denoe an individual inpu iem, i.e., a se of co-acive L1 unis (a purely spaial paern). The symbol,, is used o denoe an L2 code, i.e., a se of acive L2 unis; e.g., AB (black unis only) denoes he L2 code for he momen, [AB]. The figure also shows ha he same inpu iem, e.g., B, generally gives rise o differen L2 codes, depending on prior conex; hence, AB is no equal o ABDB, alhough hey overlap a one CM (he one a lower righ). Figure 3 graphically depics a represenaive subse of he learning ha occurs during presenaion of he sequence. All weighs, BU, TD, and H, are iniially zero. A =1, an inpu iem, A, presens a he inpu layer, L D [ABD] 4 B [ABDB]

10 Acive L1 unis send signals via heir BU weighs o L2. Assume for he momen ha some random L2 code, A, is chosen. The deails of how L2 codes are chosen will be elaboraed shorly. The BU and TD weighs beween he hree acive L1 unis comprising iem A, and he four unis of A are increased (only he increases involving he lower lef CM s winning uni are shown). The arcs from he gray L2 uni in he upper lef CM a =2 show a subse of he horizonal (emporal) learning ha would occur wihin L2 from A (which was acive a =1 and is shown as he gray unis in L2 a =2) ono AB (black unis in L2 a =2). Thus, successively acive L2 codes are chained ogeher by a emporal Hebbian learning law, Eq. 1a, in which an L2 uni acive a -1 increases is weighs ono all L2 unis acive a (excep for unis in is own CM). The learning law for he BU and TD weighs is he more familiar Hebbian rule, Eq. 1b, in which he weigh is increased if L1 uni, i, and L2 uni, j, are co-acive. w ( i, j) w 1 = (1a) 1 1 1, i, j, CM ( i) CM ( j) ( i, j), oherwise 1, i and j boh acive a w ( i, j) = (1b) w ( i, j), oherwise The model s operaion during learning is as follows. In Eq. 2, each individual L2 uni, i, compues he weighed sum, φ(i), of is horizonal (H) inpus from he L2 unis acive on he prior ime sep, -1. The analogous summaion, ψ(i), for he boom-up (BU) inpus from he currenly acive L1 unis, ϒ, is compued in Eq In Eqs. 4 and 5, hese summaions are normalized, yielding Φ(i) and Ψ(i). The φ normalizaion is possible because L2 code size is invarian and equal o Q, he number of CMs. For insance, in he model of Figure 3, all L2 codes have exacly Q = 4 acive unis. Therefore, he maximum oal horizonal inpu possible on any ime sep is Q-1 = 3 (since unis do no receive H synapses from oher unis in heir own CM). Thus, we normalize by dividing a uni s summed H inpus by Q-1. Similarly, BU inpus can be normalized because he number of acive inpu feaures on any given ime sep is assumed o be a consan, M. 4 3 This equaion is changed slighly in Sec. 4 o accommodae a slighly more complex represenaion of ime used in he compuer model. The deails will be explained in Sec The heory does no require he sric normalizaion of he number of acive inpu feaures per ime sep, bu his is no discussed in his paper. 10

11 φ ( i ) = w k i (2) k 1 ψ ( i ) Φ ( i) = = w (3) k ϒ φ ( i) Q 1 k i ψ ( i) Ψ ( i) = (5) M In Eq. 6, each L2 uni compues is own local degree of mach (suppor), (i), o he curren momen. Formally, his is implemened as muliplicaion beween he normalized versions of a uni s horizonal (H) and boom-up (BU) inpus, as indicaed in Figure 4. Noe however ha on he firs ime sep of any sequence (=1), here is no horizonal inpu presen. In his case, depends only on Ψ. This muliplicaion resuls in a sharpening of he poserior probabiliy (or likelihood) over he hypohesis space, as in he Produc of Expers model (Hinon, 1999) and consisen wih he ideas exposied by Lee & Mumford (2002). v Ψ ( i), = 1 ( i) = u v Φ ( i) Ψ ( i), > 1 Le Φ ˆ, z (depiced near he op righ of Figure 4) denoe he normalized H inpu vecor over he unis of he z h CM. Φ ˆ, z represens he emporal conex of (hisory leading up o) he curren momen,, and is a funcion of he specific se of L2 unis acive a -1. Similarly, Ψ ˆ, z (near he boom righ of Figure 4) denoes he normalized BU inpu vecor over he unis of he z h CM. Φ ˆ, z and Ψ ˆ, z are independenly exponeniaed, ypically by some small ineger (e.g., 2 or 3), o effec separae generalizaion gradiens for each influence. This furher sharpens he poserior. (4) (6) 11

12 Figure 4: Ψ ˆ, z, depics a hypoheical vecor (disribuion) of normalized BU inpus (Ψ values) o he unis of CM z a ime. The dashed ellipse represens a Ψ value of 1.0. Similarly, Φ ˆ, z depics a vecor of normalized H inpus for CM z. ˆ, z ˆ z Ψ and Φ are exponeniaed (see ex) and hen muliplied ogeher yielding a, final combined local degree of suppor disribuion, ˆ, z, over he unis of CM z. The model s abiliy o easily represen very long-disance emporal dependencies derives from he huge space of represenaions available for represening momens. In he small oy example model of Figure 3, whose layer 2 has only four CMs each wih four unis, here are 4 4 = 256 possible L2 codes, which may no seem like all ha much. However, he CM is envisioned as analogous o he corical minicolumn (Mouncasle, 1957; Peers & Sehares, 1996), or more specifically, o ha subse of a minicolumn corresponding o is principal represening cells, e.g., o a minicolumn s approximaely 30 layer 2/3 pyramidals. Furher, he model s L2 is envisioned as a pach of corex on he order of a hypercolumn, which subsumes approximaely 70 minicolumns. Such a siuaion, depiced in Figure 5, yields an asronomical number, i.e., 30 70, of unique codes. This is no o say ha all codes can be used. Because he model s learning 12

13 rule increases all weighs from one L2 code o he nex, complee sauraion of he H weigh marix, and hus oal loss of informaion, would occur long before all codes were assigned. Jus how large a fracion of he codes could be used while mainaining a given fracion of he oal sored informaion is he issue of sorage capaciy. Alhough an analyic sorage capaciy resul is no given here, empirical resuls for a simpler version of he model are given in he appendix. These resuls offer indirec evidence for subsanial capaciy, on he order of bis/synapse, as he problem is scaled o arbirarily large size. 5 Figure 5: Diagram reflecing he approximae parameers of a hypercolumn-sized pach of corex, e.g., abou 70 minicolumns where each minicolumn conains abou 30 represening cells, i.e., he layer 2/3 pyramidals. The code shown (se of black cells) is herefore one ou of a possible codes. In he nex sep of he learning algorihm, Eq. 7, he model simply finds he maximum value,, z, in each CM, z. Then, in Eq. 8, he model compues he maximum similariy, G, of he curren momen, i.e., he curren spaial inpu in he emporal conex of he sequence leading up o he curren inpu, o all previously experienced momens. Tha is, he model effecively evaluaes he likelihoods all sored hypoheses in parallel. G is normalized beween 0 and 1 and is defined as he average of he maximal 5 The possibiliy of such a vas amoun of codes (e.g., ) provides a possible answer o he binding problem. I is simply ha here are enough unique codes available o explicily represen he combinaorially large number of possible inpu configuraions for reasonably consrained inpu spaces. The same basic idea was proposed in O Reilly e al.,

14 local mach values in each CM. I can be hough of as a global mach (familiariy) signal. If G = 1, hen he curren momen is idenical o some previously experienced momen. If G is close o 0, hen he curren momen is compleely unfamiliar (novel)., z max ( i) G = (7) i CM Q z z = 1, z = (8) Q In he las phase of he learning algorihm, he model uses he global familiariy measure, G, o nonlinearly ransform he local degree of suppor values,, ino final probabiliies, ρ, of being chosen winner. Eqs. 9, 10 and 11 ogeher define how G modulaes, in a graded fashion, he characer of ha ransform. The precise parameers of hese equaions are no paricularly imporan: hey provide various means of conrolling generalizaion gradiens and he average separaion (i.e., Hamming disance) beween codes. Wha is imporan is he overall effec achieved, i.e., ha code overlap is an increasing funcion of spaioemporal similariy. a = G 1 α α + d C ( ) = a ξ i b ( i ) 1 + c ) e ρ ( i ) = k CM ( + ξ ( i ) ξ ( k ) 14 1 K (9) (10) (11) Once, he final ρ-disribuion is deermined (Eq. 11), a choice of a single winner (in each CM) is made according o i. The behavior of he model in his las phase is bes described in erms of hree differen regimes: G 1, G 0, and G somewhere in he middle. A G value close o 1 means ha here is a paricular previously experienced (sored) momen, σ*, ha is exremely similar o he curren (query) momen and herefore is causing he high G value. In his case, he model should reinsae σ* s code, σ*, wih probabiliy very close o 1.0. This is achieved by adjusing he ransform s parameers so ha i is a highly expansive nonlineariy, as shown in Figure

15 6a. This effecively causes unis wih he highes values in heir respecive CMs o win wih probabiliy close o one, and herefore σ*, as a whole, o be reinsaed wih probabiliy close o one. In conras, when G is near 0, he ransform becomes a compleely fla consan funcion (Figure 6b), which assigns equal probabiliy of winning o all unis in a module. This minimizes he average overlap of he se of winners (which is a new code) and all previously sored (known) codes. For middling G values ranging from 0 up o 1, he ransform gradually morphs from he perfecly compressive consan funcion of Figure 6b ino a highly expansive sigmoid as shown in Figure 6a. The higher he value of G, he larger he average overlap of he winning code and he closes maching sored code. Thus, his G-dependen gradual morphing of he ransform direcly yields he desired propery ha spaioemporally similar momens map o similar codes. a) G ρ b) G k ρ Figure 6: Mapping local similariy values ( values) o final probabiliies of being chosen winner (ρ values). Noe ha in he G 1 regime, he model is acually no in a learning mode. Tha is, because he choice of code depends on he produc of H signals from he previously acive code and BU signals from he inpu, exac reinsaemen of a sored code implies ha boh he prior code and he curren inpu are as hey were on he pas occasion ha is so similar o he 15

16 curren momen (i.e., causing G 1). Tha means ha here are no novel pairings, {i,j}, of acive unis, such ha uni i is acive on he prior ime sep and uni j is acive on he curren ime sep, and hus ha here are no new opporuniies for synapic increases. Thus, no learning can occur. The expeced number of novel pairings and herefore he amoun of learning ha will occur gradually increases as G drops o zero. Thus, G effecively auomaically conrols he model s movemen along he coninuum from learning o recalling/recognizing. Figure 7 depics wo hypoheical siuaions corresponding o G 1 and G 0. There is a clear maximal uni in each CM in Figure 7a. However, noe ha even in Figure 7b, here will generally sill be a maximal uni in each CM (alhough i migh no be disinguishable in his picure). In general, early in he model s learning period (life) he se of maximal unis, on each successive momen, will be idenical o some previously sored code. This is rue regardless of he value of G. This is why he model canno simply choose he maximal uni in each CM as he final winner. Doing so would lead o assigning almos all momens o a iny number of codes (even a single code), essenially losing almos all informaion abou he daa se. a) G 1.0 b) G 0.0 Figure 7: a) Hypoheical insance where G 1. b) An insance where G 0. This las poin in he discussion of he learning algorihm bridges naurally o he discussion of rerieval, i.e., recall and recogniion. For when G is near 1, i is safe, in fac, opimal, o simply choose he max uni in each CM o become acive: he winner selecion process becomes one ha can be purely local and sill be opimal. By opimal, I mean ha i 16

17 rerieves he mos similar sored momen o he curren query momen, or in oher words, rerieves he maximum likelihood hypohesis. Thus, during recall and recogniion, a each momen, he model simply acivaes he maximal uni in each CM. The recogniion algorihm consiss of Eqs. 2-6, followed by acivaing he max uni in each CM. The rerieval algorihm is slighly simpler because, following presenaion of he firs iem (i.e., he promp), only H signals are used o deermine which L2 unis become acive a each successive momen. Hence, he rerieval algorihm consiss of Eqs. 2-5, followed by Eq. 6, and hen by acivaing he maximal uni in each CM. Once he L2 code is acivaed, each L1 uni compues is oal TD inpu. Those whose TD inpu exceeds a hreshold, ypically se slighly lower han he number of L2 CMs, Q, become acive, hus replaying he original sequence a L1. ( i), 1 ( i) = Ψ = Φ ( i), > 1 An exended example follows, which illusraes wo of he model s key properies: a) spaioemporally similar momens map o similar codes, and b) he compuaion process for finding he closes maching sored momen does no depend on he number of momens (sequences) sored. This hypoheical example (Figure 8) shows he learning of he memory races for four 2-iem sequences by a small insance of he model. When he firs iem, A, of Sequence 1 presens all BU weighs are zero. All raw (ψ), and hus normalized (Ψ), BU inpu values are also zero. Thus, all local mach () values are zero, and hus, he global mach (G) is zero. Accordingly, he values are mapped hrough a consan funcion, making all unis in any given CM equally likely o win. The op panel shows a randomly chosen code, A, for iem A. Learning would occur a his poin: he BU weighs from he unis of A o he unis of A would be increased o a weigh of 1. The second iem, B, presens on he nex ime slice. Since B A = 0, none of he BU weighs from he unis of B have been increased ye. Furhermore, no H weighs have been increased ye. Since we are beyond he firs ime slice of he sequence, is compued as he produc of normalized H and BU (Φ and Ψ) inpus. However, since boh ψ and φ are zero for all L2 unis, so are all Φ, Ψ, and values, and so is G. Again, all unis end up being equally likely o win. Anoher randomly chosen code, AB, is shown for iem B (i.e., for momen [AB]). A his poin, learning occurs in he H weigh marix from he previously acive unis of A o he unis of AB, and in he BU marix from he unis of B o he unis of AB. 17 (6 )

18 Seq. 1 A AB A B Seq. 2 C CB = 0 A C =1 AB CB A C = 0 C B Seq. 3 D DB =1 A D = 2 AB DB A D =1 D B Seq. 4 E EB = 2 A E = 3 AB EB A E = 2 E B Figure 8: Hypoheical example of he firs four sequences learned by a model, illusraing he propery ha similar spaioemporal paerns (sequences) map o similar inernal represenaions (codes). Gray unis denoe inersecions wih AB. Nex, Sequence 2 presens o he model. The firs iem, C, is unfamiliar. None of C s feaures have been presen hus far. Again, G = 0; hus anoher random code, C, is chosen. Then he second iem, B, presens. Iem B has been experienced before bu in he conex of a differen predecessor. I seems reasonable ha he model should remember his novel momen, [CB], as disinc from [AB]. Thus, i should choose a novel L2 code, CB, having lile (or a leas, non-oal) overlap wih any previously chosen L2 code. While [CB] clearly has some similariy o he previously experienced momen, [AB], all values would in fac be zero in his case (because he H 18

19 inpus o all L2 cells would be zero). Thus, CB is chosen randomly: le s assume ha, by chance, i has one uni (gray cell in he lower righ CM) in common wih AB. Now Sequence 3 presens. Iem D has one feaure in common wih A (as well as one in common wih B). Various L2 unis will hus have non-zero ψ, and hus non-zero Ψ and values. However, none of he values will be close o1; hus, G will be non-zero bu closer o zero han one. The -o-ρ ransform will hus have some expansiviy and herefore cause some bias favoring he unis wih higher values, hough here will sill be significan randomness (noise) in he final choice of winners. Consequenly, we show he code, D, as having one uni in common wih A. Then iem B presens following D. This momen, [DB], has greaer spaioemporal similariy o [AB] han [CB] did. This increased similariy is manifes neurally in he uni ha is common o A and D, which means ha various L2 unis will have non-zero φ, and herefore non-zero Φ, values a his momen. Iem B is familiar: hus various L2 unis will have non-zero ψ, and hus Ψ, values. Finally, some L2 unis will have a non-zero produc,, of Φ and Ψ values; hus G will also be non-zero. Again, he ransform will have some expansiviy and hus cause some bias favoring unis wih higher values. Thus, we depic a code, DB, having wo unis in common wih AB. Finally, Sequence 4 presens. I s firs iem, E, is even more similar o A han D was (2 ou of 3 feaures in common). Wihou belaboring furher, I will jus sae ha his increased similariy leads o a code, E, having wo unis in common wih A, and ha his in urn leads o a code, EB, ha has even more overlap wih AB han did DB. Overall, his example shows how he model maps he spaioemporal similariy of momens o he (spaial) similariy of codes. This is he crucial propery of he model, which makes possible he O(1) rerieval complexiy. Figure 9 demonsraes he correcness of he model s rerieval sraegy i.e., simply picking he max uni in each CM o win rerieves he mos similar sored momen for he firs momen [A] of Sequence 1. To show his, I inroduce a varian of he G measure, he code-specific G, G( x ), which Eq. 12 defines as he average of he values for he unis comprising a specific code,, raher han he average of he maximal value in each x CM. The erm, x, z, in Eq. 12, denoes he winning uni in CM z during momen x. Each panel of Figure 9 shows one of he eigh L2 codes sored by he model during learning, along wih iem A a L1. The values, arising when iem A presens, for each of he code s unis is shown. The average of hose four values, i.e., G( x ), is shown a he op of each 19

20 panel. Since his is he fis ime sep of he rerieval even (query), he values equal he Ψ values (Eq. 7, wih exponen v = 1). ( ) z= 1 x, z x = (12) G( ) Q Q The main poins here are: a) he code-specific G, G( A), for he correc code, A, is maximal; and b) he code-specific G s for he res of he codes correlae roughly wih he similariy of he momens ha hey represen and he query momen, [A]. For example, G( A ) > G( E ) > G( D ) > G( C ), and by consrucion of he example, A A > A E > A D > A C. The code-specific G can be hough of as he relaive likelihood of he hypohesis represened by he code. G ( ) = 1.0 A G ( ) =.165 G ( ) = 0 G ( ) =.25 AB C CB = 1 =1 = 0 = 0.66 =1 =1 = 0.33 = G ( ) =.332 D G ( ) =.332 DB G ( ) =.665 E G ( ) =.082 EB =1 = 0 =1 = 0.66 = =1 = 0.33 =1 = 0.33 Figure 9: The code-specific G values (relaive likelihoods), when he query momen is [A], of all eigh momens (hypoheses) sored in he model. Noice ha he sored represenaion of [A], A, has he highes code-specific G. Figure 10 shows he code-specific G values for all eigh codes (hypoheses) when he query momen is [AB]. Thus, his is a recogniion es in which iem A was presened on he prior ime sep and iem B is now 20

21 being presened. This figure reinforces he same wo main poins as Figure 9. The correc code, AB, has he highes code-specific G, and he oher codes code-specific G s fall off roughly wih spaioemporal similariy o he query momen, [AB]. For example, G( AB ) > G( EB ) > G( DB ) > G( CB ). Again, by consrucion of he example, sim([ AB],[ AB ]) > sim([ AB],[ EB ]) > sim([ AB],[ DB ]) > sim([ AB],[ CB ]), where sim(σ i, σ j ) is he simple spaioemporal similariy meric, Eq. 13, ha depends only on he oal number of common feaures across all T ime seps. T i j = i j = 1 sim( σ, σ ) σ σ (13) G ( ) =.082 A G ( ) = 1.0 G ( ) =.082 G ( ) =.25 AB C CB = 0 = 0.33 = 1 = 1 = 1 = 1 = 0.33 = 1 G ( ) =.75 D G ( ) =.665 DB G ( ) =.082 E G ( ) =.875 EB = 1 = 1 = 1 = 0.33 = 1.33 = 1 = = 1 = 0.5 = 1 = 1 Figure 10: The code-specific G s, when he query momen is [AB] (i.e., when B is presening righ afer A presened), for all eigh sored momens. Table 1 provides he code-specific G (relaive likelihood) value for each of he eigh sored momens (hypoheses) given each of hose same momens as queries. The daa in he firs wo rows comes from Figures 9 and 10, respecively. The able demonsraes ha he same properies shown in Figures 9 and 10 hold generally for he learning se as a whole. The mos profound aspec of he model is ha he model finds he mos similar sored momen wihou explicily compuing hese code-specific G 21

22 values explicily. In fac, he model does no even have o compue G, i.e., Eq. 8, during recall and recogniion (i.e. during percepual inference). Because he way in which codes were chosen during learning maps spaioemporal similariy of momens ino degrees of code overlap and hus auomaically and implicily embeds he higher-order saisics (spaioemporal similariy srucure) of he sequence se in he model s weigh marices, simply choosing he maximal uni in each CM is guaraneed o reacivae he code of he mos similar sored momen, provided he query momen is close enough o ha sored momen. Table 1: Code-Specific G Values (Relaive Likelihoods) of all Sored Momens (hypoheses) given hose same Sored Momens as Queries Hypohesis Query A AB C CB D DB E EB A AB C CB D DB E EB Wha Table 1 does no demonsrae direcly is ha novel sequences (momens), ha are perurbaions (noisy versions) of he learning se sequences, will be mapped o he spaioemporally mos similar sored momen. However, he demonsraed code-specific G gradiens are, in some sense, he dual of ha resul. In any case, he simulaion resuls in he following secion do provide a direc demonsraion of his propery. 4. Simulaion Resuls The global familiariy, G, could be used as a signal ha ells he model wheher i is in sorage (learning) or rerieval (recall, recogniion) mode. However, he resuls repored here were produced using a simpler proocol in which he operaional mode is mandaed exernally. Thus, hese experimens have a learning phase in which he se of sequences is presened one ime each and hen a second es phase ha is eiher a recall es or a recogniion es. Weighs are frozen during he es phase. In recall ess, he model is promped wih jus he firs iem from one of he sequences ha i learned. This promp iem is also referred o as he 22

23 query, or query momen. In his case, he model reurns he remainder of he sequence. This can be hough of as a demonsraion of episodic recall of a sequence. For recall ess, we repor he recall accuracy a boh layers. To compue he accuracy, R ( ) 2, a L2, we compare he code a ime during he recall es o he code ha occurred a ime during learning, as defined in Eq. 14. E() is he number of CMs in which he uni acivaed in recall differs from he winner during learning, i.e., errors. A sequence s overall L2 accuracy is he average L2 accuracy over all ime seps (iems) of he sequence, excluding he promp ime sep. The formula for L1 recall accuracy, R ( ) 1, is slighly more complicaed since, a L1, he wo kinds of errors, deleed feaures, D(), and inruding feaures I(), mus be kep rack of separaely. The L1 code reinsaed a ime during recall is compared o he L1 code ha occurred a ime during learning, as defined in Eq. 15. C() is he number of correc feaures a L1 a ime during recall. Q E( ) R2 ( ) = (14) Q C( ) D( ) R1 ( ) = C( ) + I( ) (15) In recogniion ess, he model is presened wih an enire sequence. This query sequence could be idenical o one of he learning sequences or i could be a perurbed version of one of he learning sequences. In he former case, he model should reinsae he same sequence of L2 codes (he memory race) ha occurred during he learning rial. In he laer case, he model should acivae, on each successive momen, he L2 code corresponding o he sored momen ha is spaioemporally mos similar o he curren query momen. This demonsraes recogniion (classificaion) of a sequence. For recogniion ess, we only repor he L2 accuracy because he L1 code is supplied from he environmen on all ime seps of a recogniion rial: hus, here is no noion of an error a L1 in a recogniion es. The recogniion accuracy a L2 is also compued using Eq. 14. In his case, he comparisons are made wih he sored sequence from which he perurbed es sequence was generaed. Several parameers are common o all hree simulaion experimens. In all cases, he nework s L1 consised of 100 binary feaure deecors; he L1 and L2 unis were compleely conneced in boh he BU and TD direcion; and he L2 unis were compleely conneced via he horizonal weigh marix. All weighs are binary and iniialized o zero. 23

24 Experimen 1 Experimen 1 provides a small proof-of-concep demonsraion ha a single nework can learn a se of sequences wih single rials and subsequenly boh recall all of he individual sequences essenially perfecly and also recognize subsanially perurbed (very noisy), and hus, novel, insances of he learning se sequences. The learning se, shown in Figure 11, conained five sequences, each having five iems. Each iem consiss of en randomly seleced feaures. Thus, here are 25 unique momens. L2 consised of eigh CMs, each wih 10 unis. The recall accuracy for his se was nearly perfec; 98.18% a L2 and 100% a L1. Of he 400 reacivaions of L2 unis ha occur over he course of recalling all five sequences, only eigh errors occurred Figure 11: The five randomly creaed sequences comprising he learning se for Experimen 1. Iem (ime sep) indices are across he op. Figure 12 shows he complee learning and recogniion races, a boh L1 and L2, for he firs sequence. The op row of each race shows he L2 codes. Each row of small gray squares in an L2 grid corresponds o a CM, and each individual square in he row, o a single uni wihin ha CM. Black squares are he winners. Noice ha he race includes wo sub-seps for each iem of he sequence: e.g., iem A is acually presened wice, a sub-seps 0 and 1, iem B is presened a sub-seps 2 and 3, ec. This reflecs he compuer simulaion s propery ha signals originaing from a given inpu iem are processed up hrough all layers (in his case, here are only wo) before he nex iem is presened. Thus, on he firs sub-sep of each iem, he inpu paern iself becomes acive a L1 and on he second sub-sep, he BU signals from ha L1 code reach L2 and acivae an L2 code. Noice also ha he L2 codes also say acive for wo sub-seps, alhough hey are delayed one sub-sep from he L1 codes. This is he more complex represenaion of ime alluded o in an earlier foonoe and i enails modifying Eq. 3 o he following. The only difference is ha he 24

25 summaion is over signals arriving from L1 unis acive on he prior subsep, i.e., he se, ϒ 1, raher han he curren sub-sep. ψ ( i ) = w k i (3 ) k ϒ 1 1 (A) 2 (B) 3 (C) 4 (D) 5 (E) Iems Sub-seps Learning L2 L1 Recogniion L2 L1 Figure 12: (op) The complee learning race for Sequence 1 for boh L1 and L2. (boom) The complee recogniion race for a perurbed version of Sequence 1 which had 40% (4 ou of 10) feaures changed on each iem. The numbers below he L2 codes in he recogniion race are R 2 scores (percenages). The R 2 value of 87.5% on sub-sep 1 means ha he correc winner was reinsaed in 7 ou of he 8 CMs. See ex for furher discussion. Recogniion accuracy was also exremely good: 94.78% averaged over all five perurbed sequences. Noe ha only he final (in his case, 2 nd ) subsep of each iem is included in he compuaion of a sequence s recogniion accuracy. The inuiion here is ha as long as he recogniion L2 code racks he original (learning) L2 code on he iem imescale (as opposed o he finer sub-sep imescale), he model should be seen as demonsraing recogniion. These five sequences had a 40% perurbaion rae: i.e., 4 ou of he 10 feaures comprising each of he learning iems were randomly changed. This very subsanial degree of perurbaion can be seen by visually comparing wo L1 codes [he original learning iem (op) and is perurbed varian (boom)] as in he verically conneced boxes on he righ of in Figure 12. This race, ypical of he whole se, shows ha despie he consan exreme perurbaion (noise) on every iem of he recogniion es sequence, he model reacivaes he exac L2 code of he mos similar sored momen on each sub-sep hroughou he whole recogniion es sequence, 25

26 excep for one error on sub-sep 1 (see he verically conneced boxes on he lef; he error is in he CM represened by he op row of he L2 grid). Tha is, he model co-classifies he varian insance of Sequence 1 wih he mos similar sored sequence, which is Sequence 1. Neiher he learning algorihm (Eqs. 1-11) nor he rerieval algorihms (Eqs. 2-6 and Eqs. 2-6 ) depend on he number of sequences (or, momens) previously sored. Therefore, boh learning and rerieval ime are consan; i.e., O(1) ime complexiy. Experimen 2 The second experimen provides a larger example. In his case, here were 10 sequences, each wih 10 iems. However, his nework s L2 is much larger: 9 CMs wih 26 unis each. Again, recall accuracy was virually perfec: 99.05% a L2 and 100% a L1. The se of learning sequences, as a whole, conained 4616 bis of informaion and he nework had 95,472 weighs, yielding an informaion sorage rae of bis/synapse. While his is abou an order of magniude below he heoreical maximum for sparse binary associaive memories, hese experimens are inended as a simple proof of concep. Numerous modificaions and opimizaions are possible and will be explored in he fuure (some are briefly described in he discussion secion and appendix). Recogniion accuracy was also exremely good: 94.78% averaged over all en perurbed sequences. These sequences had a 30% perurbaion rae: i.e., 3 ou of he 10 feaures comprising each of he learning iems were randomly changed. As in Experimen 1, learning and rerieval are O(1). In his case, here are 10 x 10 = 100 sored momens whose represenaions compee a each query momen. 5. Fuure Work As presened herein, he model has a significan problem learning a large se of sequences in which many sequences sar wih he same iem or prefix of iems, e.g., all he words of some naural language. The problem is no ha he model has any inrinsic difficuly dealing wih a se of complex sequences, i.e., a sequence in which iems can recur many imes and in many differen conexs. As he example of Secion 3 shows, he choice of represenaion for an iem depends on he prior conex. Thus, here are four differen codes for iem B in ha example. In general, his allows he model o move hrough a common sae o he correc successor sae. Raher, he problem is simply ha far oo small a se of L2 codes will ever be used o represen he iniial momen of sequences in domains such as a naural language lexicon. For example, consider encoding he 26

Neural Network Model of the Backpropagation Algorithm

Neural Network Model of the Backpropagation Algorithm Neural Nework Model of he Backpropagaion Algorihm Rudolf Jakša Deparmen of Cyberneics and Arificial Inelligence Technical Universiy of Košice Lená 9, 4 Košice Slovakia jaksa@neuron.uke.sk Miroslav Karák

More information

1 Language universals

1 Language universals AS LX 500 Topics: Language Uniersals Fall 2010, Sepember 21 4a. Anisymmery 1 Language uniersals Subjec-erb agreemen and order Bach (1971) discusses wh-quesions across SO and SO languages, hypohesizing:...

More information

More Accurate Question Answering on Freebase

More Accurate Question Answering on Freebase More Accurae Quesion Answering on Freebase Hannah Bas, Elmar Haussmann Deparmen of Compuer Science Universiy of Freiburg 79110 Freiburg, Germany {bas, haussmann}@informaik.uni-freiburg.de ABSTRACT Real-world

More information

An Effiecient Approach for Resource Auto-Scaling in Cloud Environments

An Effiecient Approach for Resource Auto-Scaling in Cloud Environments Inernaional Journal of Elecrical and Compuer Engineering (IJECE) Vol. 6, No. 5, Ocober 2016, pp. 2415~2424 ISSN: 2088-8708, DOI: 10.11591/ijece.v6i5.10639 2415 An Effiecien Approach for Resource Auo-Scaling

More information

Fast Multi-task Learning for Query Spelling Correction

Fast Multi-task Learning for Query Spelling Correction Fas Muli-ask Learning for Query Spelling Correcion Xu Sun Dep. of Saisical Science Cornell Universiy Ihaca, NY 14853 xusun@cornell.edu Anshumali Shrivasava Dep. of Compuer Science Cornell Universiy Ihaca,

More information

MyLab & Mastering Business

MyLab & Mastering Business MyLab & Masering Business Efficacy Repor 2013 MyLab & Masering: Business Efficacy Repor 2013 Edied by Michelle D. Speckler 2013 Pearson MyAccouningLab, MyEconLab, MyFinanceLab, MyMarkeingLab, and MyOMLab

More information

Channel Mapping using Bidirectional Long Short-Term Memory for Dereverberation in Hands-Free Voice Controlled Devices

Channel Mapping using Bidirectional Long Short-Term Memory for Dereverberation in Hands-Free Voice Controlled Devices Z. Zhang e al.: Channel Mapping using Bidirecional Long Shor-Term Memory for Dereverberaion in Hands-Free Voice Conrolled Devices 525 Channel Mapping using Bidirecional Long Shor-Term Memory for Dereverberaion

More information

Information Propagation for informing Special Population Subgroups about New Ground Transportation Services at Airports

Information Propagation for informing Special Population Subgroups about New Ground Transportation Services at Airports Downloaded from ascelibrary.org by Basil Sephanis on 07/13/16. Copyrigh ASCE. For personal use only; all righs reserved. Informaion Propagaion for informing Special Populaion Subgroups abou New Ground

More information

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur Module 12 Machine Learning 12.1 Instructional Objective The students should understand the concept of learning systems Students should learn about different aspects of a learning system Students should

More information

Artificial Neural Networks written examination

Artificial Neural Networks written examination 1 (8) Institutionen för informationsteknologi Olle Gällmo Universitetsadjunkt Adress: Lägerhyddsvägen 2 Box 337 751 05 Uppsala Artificial Neural Networks written examination Monday, May 15, 2006 9 00-14

More information

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS Václav Kocian, Eva Volná, Michal Janošek, Martin Kotyrba University of Ostrava Department of Informatics and Computers Dvořákova 7,

More information

SARDNET: A Self-Organizing Feature Map for Sequences

SARDNET: A Self-Organizing Feature Map for Sequences SARDNET: A Self-Organizing Feature Map for Sequences Daniel L. James and Risto Miikkulainen Department of Computer Sciences The University of Texas at Austin Austin, TX 78712 dljames,risto~cs.utexas.edu

More information

Lecture 1: Machine Learning Basics

Lecture 1: Machine Learning Basics 1/69 Lecture 1: Machine Learning Basics Ali Harakeh University of Waterloo WAVE Lab ali.harakeh@uwaterloo.ca May 1, 2017 2/69 Overview 1 Learning Algorithms 2 Capacity, Overfitting, and Underfitting 3

More information

Speech Recognition at ICSI: Broadcast News and beyond

Speech Recognition at ICSI: Broadcast News and beyond Speech Recognition at ICSI: Broadcast News and beyond Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 The DARPA Broadcast News task Aspects of ICSI

More information

The Good Judgment Project: A large scale test of different methods of combining expert predictions

The Good Judgment Project: A large scale test of different methods of combining expert predictions The Good Judgment Project: A large scale test of different methods of combining expert predictions Lyle Ungar, Barb Mellors, Jon Baron, Phil Tetlock, Jaime Ramos, Sam Swift The University of Pennsylvania

More information

TD(λ) and Q-Learning Based Ludo Players

TD(λ) and Q-Learning Based Ludo Players TD(λ) and Q-Learning Based Ludo Players Majed Alhajry, Faisal Alvi, Member, IEEE and Moataz Ahmed Abstract Reinforcement learning is a popular machine learning technique whose inherent self-learning ability

More information

Comparison of network inference packages and methods for multiple networks inference

Comparison of network inference packages and methods for multiple networks inference Comparison of network inference packages and methods for multiple networks inference Nathalie Villa-Vialaneix http://www.nathalievilla.org nathalie.villa@univ-paris1.fr 1ères Rencontres R - BoRdeaux, 3

More information

Axiom 2013 Team Description Paper

Axiom 2013 Team Description Paper Axiom 2013 Team Description Paper Mohammad Ghazanfari, S Omid Shirkhorshidi, Farbod Samsamipour, Hossein Rahmatizadeh Zagheli, Mohammad Mahdavi, Payam Mohajeri, S Abbas Alamolhoda Robotics Scientific Association

More information

Exploration. CS : Deep Reinforcement Learning Sergey Levine

Exploration. CS : Deep Reinforcement Learning Sergey Levine Exploration CS 294-112: Deep Reinforcement Learning Sergey Levine Class Notes 1. Homework 4 due on Wednesday 2. Project proposal feedback sent Today s Lecture 1. What is exploration? Why is it a problem?

More information

Lecture 10: Reinforcement Learning

Lecture 10: Reinforcement Learning Lecture 1: Reinforcement Learning Cognitive Systems II - Machine Learning SS 25 Part III: Learning Programs and Strategies Q Learning, Dynamic Programming Lecture 1: Reinforcement Learning p. Motivation

More information

Transfer Learning Action Models by Measuring the Similarity of Different Domains

Transfer Learning Action Models by Measuring the Similarity of Different Domains Transfer Learning Action Models by Measuring the Similarity of Different Domains Hankui Zhuo 1, Qiang Yang 2, and Lei Li 1 1 Software Research Institute, Sun Yat-sen University, Guangzhou, China. zhuohank@gmail.com,lnslilei@mail.sysu.edu.cn

More information

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria

FUZZY EXPERT. Dr. Kasim M. Al-Aubidy. Philadelphia University. Computer Eng. Dept February 2002 University of Damascus-Syria FUZZY EXPERT SYSTEMS 16-18 18 February 2002 University of Damascus-Syria Dr. Kasim M. Al-Aubidy Computer Eng. Dept. Philadelphia University What is Expert Systems? ES are computer programs that emulate

More information

Learning Methods in Multilingual Speech Recognition

Learning Methods in Multilingual Speech Recognition Learning Methods in Multilingual Speech Recognition Hui Lin Department of Electrical Engineering University of Washington Seattle, WA 98125 linhui@u.washington.edu Li Deng, Jasha Droppo, Dong Yu, and Alex

More information

Active Learning. Yingyu Liang Computer Sciences 760 Fall

Active Learning. Yingyu Liang Computer Sciences 760 Fall Active Learning Yingyu Liang Computer Sciences 760 Fall 2017 http://pages.cs.wisc.edu/~yliang/cs760/ Some of the slides in these lectures have been adapted/borrowed from materials developed by Mark Craven,

More information

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition Chapter 2: The Representation of Knowledge Expert Systems: Principles and Programming, Fourth Edition Objectives Introduce the study of logic Learn the difference between formal logic and informal logic

More information

An Introduction to Simio for Beginners

An Introduction to Simio for Beginners An Introduction to Simio for Beginners C. Dennis Pegden, Ph.D. This white paper is intended to introduce Simio to a user new to simulation. It is intended for the manufacturing engineer, hospital quality

More information

Reinforcement Learning by Comparing Immediate Reward

Reinforcement Learning by Comparing Immediate Reward Reinforcement Learning by Comparing Immediate Reward Punit Pandey DeepshikhaPandey Dr. Shishir Kumar Abstract This paper introduces an approach to Reinforcement Learning Algorithm by comparing their immediate

More information

ACTIVITY: Comparing Combination Locks

ACTIVITY: Comparing Combination Locks 5.4 Compound Events outcomes of one or more events? ow can you find the number of possible ACIVIY: Comparing Combination Locks Work with a partner. You are buying a combination lock. You have three choices.

More information

A Neural Network GUI Tested on Text-To-Phoneme Mapping

A Neural Network GUI Tested on Text-To-Phoneme Mapping A Neural Network GUI Tested on Text-To-Phoneme Mapping MAARTEN TROMPPER Universiteit Utrecht m.f.a.trompper@students.uu.nl Abstract Text-to-phoneme (T2P) mapping is a necessary step in any speech synthesis

More information

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1 Notes on The Sciences of the Artificial Adapted from a shorter document written for course 17-652 (Deciding What to Design) 1 Ali Almossawi December 29, 2005 1 Introduction The Sciences of the Artificial

More information

CONCEPT MAPS AS A DEVICE FOR LEARNING DATABASE CONCEPTS

CONCEPT MAPS AS A DEVICE FOR LEARNING DATABASE CONCEPTS CONCEPT MAPS AS A DEVICE FOR LEARNING DATABASE CONCEPTS Pirjo Moen Department of Computer Science P.O. Box 68 FI-00014 University of Helsinki pirjo.moen@cs.helsinki.fi http://www.cs.helsinki.fi/pirjo.moen

More information

Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C

Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C Numeracy Medium term plan: Summer Term Level 2C/2B Year 2 Level 2A/3C Using and applying mathematics objectives (Problem solving, Communicating and Reasoning) Select the maths to use in some classroom

More information

A Version Space Approach to Learning Context-free Grammars

A Version Space Approach to Learning Context-free Grammars Machine Learning 2: 39~74, 1987 1987 Kluwer Academic Publishers, Boston - Manufactured in The Netherlands A Version Space Approach to Learning Context-free Grammars KURT VANLEHN (VANLEHN@A.PSY.CMU.EDU)

More information

Evolutive Neural Net Fuzzy Filtering: Basic Description

Evolutive Neural Net Fuzzy Filtering: Basic Description Journal of Intelligent Learning Systems and Applications, 2010, 2: 12-18 doi:10.4236/jilsa.2010.21002 Published Online February 2010 (http://www.scirp.org/journal/jilsa) Evolutive Neural Net Fuzzy Filtering:

More information

Rule Learning With Negation: Issues Regarding Effectiveness

Rule Learning With Negation: Issues Regarding Effectiveness Rule Learning With Negation: Issues Regarding Effectiveness S. Chua, F. Coenen, G. Malcolm University of Liverpool Department of Computer Science, Ashton Building, Ashton Street, L69 3BX Liverpool, United

More information

Human Emotion Recognition From Speech

Human Emotion Recognition From Speech RESEARCH ARTICLE OPEN ACCESS Human Emotion Recognition From Speech Miss. Aparna P. Wanare*, Prof. Shankar N. Dandare *(Department of Electronics & Telecommunication Engineering, Sant Gadge Baba Amravati

More information

MYCIN. The MYCIN Task

MYCIN. The MYCIN Task MYCIN Developed at Stanford University in 1972 Regarded as the first true expert system Assists physicians in the treatment of blood infections Many revisions and extensions over the years The MYCIN Task

More information

Python Machine Learning

Python Machine Learning Python Machine Learning Unlock deeper insights into machine learning with this vital guide to cuttingedge predictive analytics Sebastian Raschka [ PUBLISHING 1 open source I community experience distilled

More information

School Size and the Quality of Teaching and Learning

School Size and the Quality of Teaching and Learning School Size and the Quality of Teaching and Learning An Analysis of Relationships between School Size and Assessments of Factors Related to the Quality of Teaching and Learning in Primary Schools Undertaken

More information

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks 1 Tzu-Hsuan Yang, 2 Tzu-Hsuan Tseng, and 3 Chia-Ping Chen Department of Computer Science and Engineering

More information

THE ROLE OF TOOL AND TEACHER MEDIATIONS IN THE CONSTRUCTION OF MEANINGS FOR REFLECTION

THE ROLE OF TOOL AND TEACHER MEDIATIONS IN THE CONSTRUCTION OF MEANINGS FOR REFLECTION THE ROLE OF TOOL AND TEACHER MEDIATIONS IN THE CONSTRUCTION OF MEANINGS FOR REFLECTION Lulu Healy Programa de Estudos Pós-Graduados em Educação Matemática, PUC, São Paulo ABSTRACT This article reports

More information

Seminar - Organic Computing

Seminar - Organic Computing Seminar - Organic Computing Self-Organisation of OC-Systems Markus Franke 25.01.2006 Typeset by FoilTEX Timetable 1. Overview 2. Characteristics of SO-Systems 3. Concern with Nature 4. Design-Concepts

More information

CS Machine Learning

CS Machine Learning CS 478 - Machine Learning Projects Data Representation Basic testing and evaluation schemes CS 478 Data and Testing 1 Programming Issues l Program in any platform you want l Realize that you will be doing

More information

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge Innov High Educ (2009) 34:93 103 DOI 10.1007/s10755-009-9095-2 Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge Phyllis Blumberg Published online: 3 February

More information

Page 1 of 11. Curriculum Map: Grade 4 Math Course: Math 4 Sub-topic: General. Grade(s): None specified

Page 1 of 11. Curriculum Map: Grade 4 Math Course: Math 4 Sub-topic: General. Grade(s): None specified Curriculum Map: Grade 4 Math Course: Math 4 Sub-topic: General Grade(s): None specified Unit: Creating a Community of Mathematical Thinkers Timeline: Week 1 The purpose of the Establishing a Community

More information

Syntactic systematicity in sentence processing with a recurrent self-organizing network

Syntactic systematicity in sentence processing with a recurrent self-organizing network Syntactic systematicity in sentence processing with a recurrent self-organizing network Igor Farkaš,1 Department of Applied Informatics, Comenius University Mlynská dolina, 842 48 Bratislava, Slovak Republic

More information

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words, A Language-Independent, Data-Oriented Architecture for Grapheme-to-Phoneme Conversion Walter Daelemans and Antal van den Bosch Proceedings ESCA-IEEE speech synthesis conference, New York, September 1994

More information

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE

Course Outline. Course Grading. Where to go for help. Academic Integrity. EE-589 Introduction to Neural Networks NN 1 EE EE-589 Introduction to Neural Assistant Prof. Dr. Turgay IBRIKCI Room # 305 (322) 338 6868 / 139 Wensdays 9:00-12:00 Course Outline The course is divided in two parts: theory and practice. 1. Theory covers

More information

The Evolution of Random Phenomena

The Evolution of Random Phenomena The Evolution of Random Phenomena A Look at Markov Chains Glen Wang glenw@uchicago.edu Splash! Chicago: Winter Cascade 2012 Lecture 1: What is Randomness? What is randomness? Can you think of some examples

More information

Julia Smith. Effective Classroom Approaches to.

Julia Smith. Effective Classroom Approaches to. Julia Smith @tessmaths Effective Classroom Approaches to GCSE Maths resits julia.smith@writtle.ac.uk Agenda The context of GCSE resit in a post-16 setting An overview of the new GCSE Key features of a

More information

Cross Language Information Retrieval

Cross Language Information Retrieval Cross Language Information Retrieval RAFFAELLA BERNARDI UNIVERSITÀ DEGLI STUDI DI TRENTO P.ZZA VENEZIA, ROOM: 2.05, E-MAIL: BERNARDI@DISI.UNITN.IT Contents 1 Acknowledgment.............................................

More information

Longitudinal Analysis of the Effectiveness of DCPS Teachers

Longitudinal Analysis of the Effectiveness of DCPS Teachers F I N A L R E P O R T Longitudinal Analysis of the Effectiveness of DCPS Teachers July 8, 2014 Elias Walsh Dallas Dotter Submitted to: DC Education Consortium for Research and Evaluation School of Education

More information

On-Line Data Analytics

On-Line Data Analytics International Journal of Computer Applications in Engineering Sciences [VOL I, ISSUE III, SEPTEMBER 2011] [ISSN: 2231-4946] On-Line Data Analytics Yugandhar Vemulapalli #, Devarapalli Raghu *, Raja Jacob

More information

IMGD Technical Game Development I: Iterative Development Techniques. by Robert W. Lindeman

IMGD Technical Game Development I: Iterative Development Techniques. by Robert W. Lindeman IMGD 3000 - Technical Game Development I: Iterative Development Techniques by Robert W. Lindeman gogo@wpi.edu Motivation The last thing you want to do is write critical code near the end of a project Induces

More information

Given a directed graph G =(N A), where N is a set of m nodes and A. destination node, implying a direction for ow to follow. Arcs have limitations

Given a directed graph G =(N A), where N is a set of m nodes and A. destination node, implying a direction for ow to follow. Arcs have limitations 4 Interior point algorithms for network ow problems Mauricio G.C. Resende AT&T Bell Laboratories, Murray Hill, NJ 07974-2070 USA Panos M. Pardalos The University of Florida, Gainesville, FL 32611-6595

More information

Linking the Ohio State Assessments to NWEA MAP Growth Tests *

Linking the Ohio State Assessments to NWEA MAP Growth Tests * Linking the Ohio State Assessments to NWEA MAP Growth Tests * *As of June 2017 Measures of Academic Progress (MAP ) is known as MAP Growth. August 2016 Introduction Northwest Evaluation Association (NWEA

More information

Lahore University of Management Sciences. FINN 321 Econometrics Fall Semester 2017

Lahore University of Management Sciences. FINN 321 Econometrics Fall Semester 2017 Instructor Syed Zahid Ali Room No. 247 Economics Wing First Floor Office Hours Email szahid@lums.edu.pk Telephone Ext. 8074 Secretary/TA TA Office Hours Course URL (if any) Suraj.lums.edu.pk FINN 321 Econometrics

More information

Entrepreneurial Discovery and the Demmert/Klein Experiment: Additional Evidence from Germany

Entrepreneurial Discovery and the Demmert/Klein Experiment: Additional Evidence from Germany Entrepreneurial Discovery and the Demmert/Klein Experiment: Additional Evidence from Germany Jana Kitzmann and Dirk Schiereck, Endowed Chair for Banking and Finance, EUROPEAN BUSINESS SCHOOL, International

More information

University of Waterloo School of Accountancy. AFM 102: Introductory Management Accounting. Fall Term 2004: Section 4

University of Waterloo School of Accountancy. AFM 102: Introductory Management Accounting. Fall Term 2004: Section 4 University of Waterloo School of Accountancy AFM 102: Introductory Management Accounting Fall Term 2004: Section 4 Instructor: Alan Webb Office: HH 289A / BFG 2120 B (after October 1) Phone: 888-4567 ext.

More information

Cued Recall From Image and Sentence Memory: A Shift From Episodic to Identical Elements Representation

Cued Recall From Image and Sentence Memory: A Shift From Episodic to Identical Elements Representation Journal of Experimental Psychology: Learning, Memory, and Cognition 2006, Vol. 32, No. 4, 734 748 Copyright 2006 by the American Psychological Association 0278-7393/06/$12.00 DOI: 10.1037/0278-7393.32.4.734

More information

Grade 6: Correlated to AGS Basic Math Skills

Grade 6: Correlated to AGS Basic Math Skills Grade 6: Correlated to AGS Basic Math Skills Grade 6: Standard 1 Number Sense Students compare and order positive and negative integers, decimals, fractions, and mixed numbers. They find multiples and

More information

A Case-Based Approach To Imitation Learning in Robotic Agents

A Case-Based Approach To Imitation Learning in Robotic Agents A Case-Based Approach To Imitation Learning in Robotic Agents Tesca Fitzgerald, Ashok Goel School of Interactive Computing Georgia Institute of Technology, Atlanta, GA 30332, USA {tesca.fitzgerald,goel}@cc.gatech.edu

More information

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Chinese Language Parsing with Maximum-Entropy-Inspired Parser Chinese Language Parsing with Maximum-Entropy-Inspired Parser Heng Lian Brown University Abstract The Chinese language has many special characteristics that make parsing difficult. The performance of state-of-the-art

More information

WHEN THERE IS A mismatch between the acoustic

WHEN THERE IS A mismatch between the acoustic 808 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 14, NO. 3, MAY 2006 Optimization of Temporal Filters for Constructing Robust Features in Speech Recognition Jeih-Weih Hung, Member,

More information

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM

ISFA2008U_120 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM Proceedings of 28 ISFA 28 International Symposium on Flexible Automation Atlanta, GA, USA June 23-26, 28 ISFA28U_12 A SCHEDULING REINFORCEMENT LEARNING ALGORITHM Amit Gil, Helman Stern, Yael Edan, and

More information

Year 4 National Curriculum requirements

Year 4 National Curriculum requirements Year National Curriculum requirements Pupils should be taught to develop a range of personal strategies for learning new and irregular words* develop a range of personal strategies for spelling at the

More information

School Competition and Efficiency with Publicly Funded Catholic Schools David Card, Martin D. Dooley, and A. Abigail Payne

School Competition and Efficiency with Publicly Funded Catholic Schools David Card, Martin D. Dooley, and A. Abigail Payne School Competition and Efficiency with Publicly Funded Catholic Schools David Card, Martin D. Dooley, and A. Abigail Payne Web Appendix See paper for references to Appendix Appendix 1: Multiple Schools

More information

Learning to Rank with Selection Bias in Personal Search

Learning to Rank with Selection Bias in Personal Search Learning to Rank with Selection Bias in Personal Search Xuanhui Wang, Michael Bendersky, Donald Metzler, Marc Najork Google Inc. Mountain View, CA 94043 {xuanhui, bemike, metzler, najork}@google.com ABSTRACT

More information

Evolution of Symbolisation in Chimpanzees and Neural Nets

Evolution of Symbolisation in Chimpanzees and Neural Nets Evolution of Symbolisation in Chimpanzees and Neural Nets Angelo Cangelosi Centre for Neural and Adaptive Systems University of Plymouth (UK) a.cangelosi@plymouth.ac.uk Introduction Animal communication

More information

Session 2B From understanding perspectives to informing public policy the potential and challenges for Q findings to inform survey design

Session 2B From understanding perspectives to informing public policy the potential and challenges for Q findings to inform survey design Session 2B From understanding perspectives to informing public policy the potential and challenges for Q findings to inform survey design Paper #3 Five Q-to-survey approaches: did they work? Job van Exel

More information

Welcome to ACT Brain Boot Camp

Welcome to ACT Brain Boot Camp Welcome to ACT Brain Boot Camp 9:30 am - 9:45 am Basics (in every room) 9:45 am - 10:15 am Breakout Session #1 ACT Math: Adame ACT Science: Moreno ACT Reading: Campbell ACT English: Lee 10:20 am - 10:50

More information

Rule Learning with Negation: Issues Regarding Effectiveness

Rule Learning with Negation: Issues Regarding Effectiveness Rule Learning with Negation: Issues Regarding Effectiveness Stephanie Chua, Frans Coenen, and Grant Malcolm University of Liverpool Department of Computer Science, Ashton Building, Ashton Street, L69 3BX

More information

Learning Methods for Fuzzy Systems

Learning Methods for Fuzzy Systems Learning Methods for Fuzzy Systems Rudolf Kruse and Andreas Nürnberger Department of Computer Science, University of Magdeburg Universitätsplatz, D-396 Magdeburg, Germany Phone : +49.39.67.876, Fax : +49.39.67.8

More information

Measurement. When Smaller Is Better. Activity:

Measurement. When Smaller Is Better. Activity: Measurement Activity: TEKS: When Smaller Is Better (6.8) Measurement. The student solves application problems involving estimation and measurement of length, area, time, temperature, volume, weight, and

More information

Extending Place Value with Whole Numbers to 1,000,000

Extending Place Value with Whole Numbers to 1,000,000 Grade 4 Mathematics, Quarter 1, Unit 1.1 Extending Place Value with Whole Numbers to 1,000,000 Overview Number of Instructional Days: 10 (1 day = 45 minutes) Content to Be Learned Recognize that a digit

More information

Model Ensemble for Click Prediction in Bing Search Ads

Model Ensemble for Click Prediction in Bing Search Ads Model Ensemble for Click Prediction in Bing Search Ads Xiaoliang Ling Microsoft Bing xiaoling@microsoft.com Hucheng Zhou Microsoft Research huzho@microsoft.com Weiwei Deng Microsoft Bing dedeng@microsoft.com

More information

Loughton School s curriculum evening. 28 th February 2017

Loughton School s curriculum evening. 28 th February 2017 Loughton School s curriculum evening 28 th February 2017 Aims of this session Share our approach to teaching writing, reading, SPaG and maths. Share resources, ideas and strategies to support children's

More information

MTH 141 Calculus 1 Syllabus Spring 2017

MTH 141 Calculus 1 Syllabus Spring 2017 Instructor: Section/Meets Office Hrs: Textbook: Calculus: Single Variable, by Hughes-Hallet et al, 6th ed., Wiley. Also needed: access code to WileyPlus (included in new books) Calculator: Not required,

More information

Probabilistic Latent Semantic Analysis

Probabilistic Latent Semantic Analysis Probabilistic Latent Semantic Analysis Thomas Hofmann Presentation by Ioannis Pavlopoulos & Andreas Damianou for the course of Data Mining & Exploration 1 Outline Latent Semantic Analysis o Need o Overview

More information

Clouds = Heavy Sidewalk = Wet. davinci V2.1 alpha3

Clouds = Heavy Sidewalk = Wet. davinci V2.1 alpha3 Identifying and Handling Structural Incompleteness for Validation of Probabilistic Knowledge-Bases Eugene Santos Jr. Dept. of Comp. Sci. & Eng. University of Connecticut Storrs, CT 06269-3155 eugene@cse.uconn.edu

More information

Detecting English-French Cognates Using Orthographic Edit Distance

Detecting English-French Cognates Using Orthographic Edit Distance Detecting English-French Cognates Using Orthographic Edit Distance Qiongkai Xu 1,2, Albert Chen 1, Chang i 1 1 The Australian National University, College of Engineering and Computer Science 2 National

More information

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016 AGENDA Advanced Learning Theories Alejandra J. Magana, Ph.D. admagana@purdue.edu Introduction to Learning Theories Role of Learning Theories and Frameworks Learning Design Research Design Dual Coding Theory

More information

Monitoring Metacognitive abilities in children: A comparison of children between the ages of 5 to 7 years and 8 to 11 years

Monitoring Metacognitive abilities in children: A comparison of children between the ages of 5 to 7 years and 8 to 11 years Monitoring Metacognitive abilities in children: A comparison of children between the ages of 5 to 7 years and 8 to 11 years Abstract Takang K. Tabe Department of Educational Psychology, University of Buea

More information

College Pricing. Ben Johnson. April 30, Abstract. Colleges in the United States price discriminate based on student characteristics

College Pricing. Ben Johnson. April 30, Abstract. Colleges in the United States price discriminate based on student characteristics College Pricing Ben Johnson April 30, 2012 Abstract Colleges in the United States price discriminate based on student characteristics such as ability and income. This paper develops a model of college

More information

What the National Curriculum requires in reading at Y5 and Y6

What the National Curriculum requires in reading at Y5 and Y6 What the National Curriculum requires in reading at Y5 and Y6 Word reading apply their growing knowledge of root words, prefixes and suffixes (morphology and etymology), as listed in Appendix 1 of the

More information

***** Article in press in Neural Networks ***** BOTTOM-UP LEARNING OF EXPLICIT KNOWLEDGE USING A BAYESIAN ALGORITHM AND A NEW HEBBIAN LEARNING RULE

***** Article in press in Neural Networks ***** BOTTOM-UP LEARNING OF EXPLICIT KNOWLEDGE USING A BAYESIAN ALGORITHM AND A NEW HEBBIAN LEARNING RULE Bottom-up learning of explicit knowledge 1 ***** Article in press in Neural Networks ***** BOTTOM-UP LEARNING OF EXPLICIT KNOWLEDGE USING A BAYESIAN ALGORITHM AND A NEW HEBBIAN LEARNING RULE Sébastien

More information

College Pricing and Income Inequality

College Pricing and Income Inequality College Pricing and Income Inequality Zhifeng Cai U of Minnesota, Rutgers University, and FRB Minneapolis Jonathan Heathcote FRB Minneapolis NBER Income Distribution, July 20, 2017 The views expressed

More information

Functional Skills Mathematics Level 2 assessment

Functional Skills Mathematics Level 2 assessment Functional Skills Mathematics Level 2 assessment www.cityandguilds.com September 2015 Version 1.0 Marking scheme ONLINE V2 Level 2 Sample Paper 4 Mark Represent Analyse Interpret Open Fixed S1Q1 3 3 0

More information

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

OCR for Arabic using SIFT Descriptors With Online Failure Prediction OCR for Arabic using SIFT Descriptors With Online Failure Prediction Andrey Stolyarenko, Nachum Dershowitz The Blavatnik School of Computer Science Tel Aviv University Tel Aviv, Israel Email: stloyare@tau.ac.il,

More information

Digital Fabrication and Aunt Sarah: Enabling Quadratic Explorations via Technology. Michael L. Connell University of Houston - Downtown

Digital Fabrication and Aunt Sarah: Enabling Quadratic Explorations via Technology. Michael L. Connell University of Houston - Downtown Digital Fabrication and Aunt Sarah: Enabling Quadratic Explorations via Technology Michael L. Connell University of Houston - Downtown Sergei Abramovich State University of New York at Potsdam Introduction

More information

Dublin City Schools Mathematics Graded Course of Study GRADE 4

Dublin City Schools Mathematics Graded Course of Study GRADE 4 I. Content Standard: Number, Number Sense and Operations Standard Students demonstrate number sense, including an understanding of number systems and reasonable estimates using paper and pencil, technology-supported

More information

A General Class of Noncontext Free Grammars Generating Context Free Languages

A General Class of Noncontext Free Grammars Generating Context Free Languages INFORMATION AND CONTROL 43, 187-194 (1979) A General Class of Noncontext Free Grammars Generating Context Free Languages SARWAN K. AGGARWAL Boeing Wichita Company, Wichita, Kansas 67210 AND JAMES A. HEINEN

More information

Kelli Allen. Vicki Nieter. Jeanna Scheve. Foreword by Gregory J. Kaiser

Kelli Allen. Vicki Nieter. Jeanna Scheve. Foreword by Gregory J. Kaiser Kelli Allen Jeanna Scheve Vicki Nieter Foreword by Gregory J. Kaiser Table of Contents Foreword........................................... 7 Introduction........................................ 9 Learning

More information

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science

Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science Proposal of Pattern Recognition as a necessary and sufficient principle to Cognitive Science Gilberto de Paiva Sao Paulo Brazil (May 2011) gilbertodpaiva@gmail.com Abstract. Despite the prevalence of the

More information

Artificial Neural Networks

Artificial Neural Networks Artificial Neural Networks Andres Chavez Math 382/L T/Th 2:00-3:40 April 13, 2010 Chavez2 Abstract The main interest of this paper is Artificial Neural Networks (ANNs). A brief history of the development

More information

How do adults reason about their opponent? Typologies of players in a turn-taking game

How do adults reason about their opponent? Typologies of players in a turn-taking game How do adults reason about their opponent? Typologies of players in a turn-taking game Tamoghna Halder (thaldera@gmail.com) Indian Statistical Institute, Kolkata, India Khyati Sharma (khyati.sharma27@gmail.com)

More information

arxiv: v1 [cs.cl] 2 Apr 2017

arxiv: v1 [cs.cl] 2 Apr 2017 Word-Alignment-Based Segment-Level Machine Translation Evaluation using Word Embeddings Junki Matsuo and Mamoru Komachi Graduate School of System Design, Tokyo Metropolitan University, Japan matsuo-junki@ed.tmu.ac.jp,

More information

An Evaluation of the Interactive-Activation Model Using Masked Partial-Word Priming. Jason R. Perry. University of Western Ontario. Stephen J.

An Evaluation of the Interactive-Activation Model Using Masked Partial-Word Priming. Jason R. Perry. University of Western Ontario. Stephen J. An Evaluation of the Interactive-Activation Model Using Masked Partial-Word Priming Jason R. Perry University of Western Ontario Stephen J. Lupker University of Western Ontario Colin J. Davis Royal Holloway

More information

12- A whirlwind tour of statistics

12- A whirlwind tour of statistics CyLab HT 05-436 / 05-836 / 08-534 / 08-734 / 19-534 / 19-734 Usable Privacy and Security TP :// C DU February 22, 2016 y & Secu rivac rity P le ratory bo La Lujo Bauer, Nicolas Christin, and Abby Marsh

More information

Understanding and Supporting Dyslexia Godstone Village School. January 2017

Understanding and Supporting Dyslexia Godstone Village School. January 2017 Understanding and Supporting Dyslexia Godstone Village School January 2017 By then end of the session I will: Have a greater understanding of Dyslexia and the ways in which children can be affected by

More information