ÌìȻ˵»°´¦ÖÃÉî¶È½ø½¨ÍøÂçÉè¼Æ²½Öè

2019.05.20

Ͷ¸å£ºÑîÐãÀö²¿ÃÅ£ºÍ¨Ñ¶ÓëÐÅÏ¢¹¤³ÌѧԺä¯ÀÀ´ÎÊý£º

»î¶¯ÐÅÏ¢

¹¦·ò£º 2019Äê05ÔÂ22ÈÕ 10:30

µØÖ·£º У±¾²¿¶«ÇøÏèÓ¢´óÂ¥T706ÊÒ

Ðн¡½²Ì³Ñ§Êõ½²×ù

µÚ393ÆÚ

¹¦·ò:   2019Äê5ÔÂ22ÈÕ£¨ÖÜÈý£©ÉÏÎç10£º30-11: 30

µØÖ·:   У±¾²¿¶«ÇøÏèÓ¢´óÂ¥T706ÊÒ

½²×ù:   ÌìȻ˵»°´¦ÖÃÉî¶È½ø½¨ÍøÂçÉè¼Æ²½Öè

 

Ñݽ²Õß: ·ðÂÞÀï´ï´óѧ  Îâ´óÅô½ÌÊÚ  IEEE Fellow

Ñݽ²Õß¼ò½é£ºDapeng Oliver Wu received Ph.D. in Electrical and Computer Engineering from Carnegie Mellon University, Pittsburgh, PA, in 2003. Since 2003, he has been on the faculty of Electrical and Computer Engineering Department at University of Florida, Gainesville, FL, where he is currently Professor.  His research interests are in the areas of networking, communications, video coding, image processing, computer vision, signal processing, and machine learning. He received University of Florida Term Professorship Award in 2017, University of Florida Research Foundation Professorship Award in 2009, AFOSR Young Investigator Program (YIP) Award in 2009, ONR Young Investigator Program (YIP) Award in 2008, NSF CAREER award in 2007, the IEEE Circuits and Systems for Video Technology (CSVT) Transactions Best Paper Award for Year 2001, the Best Paper Award in GLOBECOM 2011, and the Best Paper Award in QShine 2006. 

½²×ùÌáÒª£ºIn this talk, I present a new approach to the design of deep networks for natural language processing (NLP), based on the general technique of Tensor Product Representations (TPRs) for encoding and processing symbol structures in distributed neural networks. A network architecture--- the Tensor Product Generation Network (TPGN) --- is proposed which is capable in principle of carrying out TPR computation, but which uses unconstrained deep learning to design its internal representations. Instantiated in a model for image-caption generation, TPGN outperforms LSTM baselines when evaluated on the COCO dataset. The TPR-capable structure enables interpretation of internal representations and operations, which prove to contain considerable grammatical content. Our caption-generation model can be interpreted as generating sequences of grammatical categories and retrieving words by their categories from a plan encoded as a distributed representation.

Ô¼ÇëÈË£ºÐ±¦GGͨѶÓëÐÅÏ¢¹¤³ÌѧԺ ÉòÀñȨ×êÑÐÔ±

Ó­½Ó¿í´óÀÏʦºÍѧÉú²ÎÓ룡


¡¾ÍøÕ¾µØÍ¼¡¿