Заключение
633
562. Opper, M. and Archambeau, C. (2009). The variational Gaussian approximation re-
visited.
Neural computation, 21(3), 786–792.
563. Oquab, M., Bottou, L., Laptev, I., and Sivic, J. (2014). Learning and transferring
mid-level image representations using convolutional neural networks. In Computer
Vision and Pattern Recognition (CVPR), 2014 IEEE Conference on, pages 1717–
1724. IEEE.
564. Osindero, S.
and Hinton, G. E. (2008). Modeling image patches with a directed
hie rarchy of Markov random fields. In J. Platt, D. Koller, Y. Singer, and S. Ro weis,
editors, Advances in Neural Information Processing Systems 20 (NIPS’07), pa-
ges 1121–1128, Cambridge, MA. MIT Press.
565. Ovid and Martin, C. (2004). Metamorphoses. W.W. Norton.
566. Paccanaro, A. and Hinton, G. E. (2000). Extracting distributed representations of
concepts and relations from positive and negative propositions.
In International
Joint Conference on Neural Networks (IJCNN), Como, Italy. IEEE, New York.
567. Paine, T. L., Khorrami, P., Han, W., and Huang, T. S. (2014). An analysis of unsuper-
vised pre-training in light of recent advances. arXiv preprint arXiv:1412.6597.
568. Palatucci, M., Pomerleau, D., Hinton, G. E., and Mitchell, T. M. (2009). Zero-shot
learning with semantic output codes. In Y. Bengio, D. Schuurmans, J. D. Lafferty,
C. K. I. Williams, and A. Culotta, editors, Advances in Neural Information Process-
ing Systems 22, pages 1410–1418. Curran Associates, Inc.
569. Parker, D. B. (1985). Learning-logic. Technical Report TR-47, Center for Comp. Re-
search in Economics and Management Sci., MIT.
570. Pascanu, R., Mikolov, T., and Bengio, Y. (2013). On the difficulty of training recur-
rent neural networks. In ICML’2013.
571. Pascanu, R., G
ü
l
ç
ehre,
Ç
., Cho, K., and Bengio, Y. (2014a). How to construct deep
recurrent neural networks. In ICLR’2014.
572. Pascanu, R., Montufar, G., and Bengio, Y. (2014b). On the number of inference re-
gions of deep feed forward networks with piece-wise linear activations. In ICLR’2014.
573. Pati, Y., Rezaiifar, R., and Krishnaprasad, P. (1993). Orthogonal matching pursuit:
Recursive function approximation with applications to wavelet decomposition. In
Proceedings of the 27 th Annual Asilomar Conference on Signals, Systems, and Com-
puters, pages 40–44.
574. Pearl, J. (1985). Bayesian networks: A model of self-activated memory for evidential
reasoning. In Proceedings of the 7th Conference of the Cognitive Science Society,
University of California, Irvine, pages 329–334.
575. Pearl, J. (1988). Probabilistic Reasoning in Intelligent Systems: Networks of Plau-
sible Inference. Morgan Kaufmann.
576. Perron, O. (1907). Zur theorie der matrices. Mathematische Annalen, 64(2), 248–
263.
577.
Petersen, K. B. and Pedersen, M. S. (2006). The matrix cookbook. Version 20051003.
578. Peterson, G. B. (2004). A day of great illumination: B. F. Skinner’s discovery of shap-
ing. Journal of the Experimental Analysis of Behavior, 82(3), 317–328.
579. Pham, D.-T., Garat, P., and Jutten, C. (1992). Separation of a mixture of independent
sources through a maximum likelihood approach. In EUSIPCO, pages 771–774.
580. Pham, P.-H., Jelaca, D., Farabet, C., Martini, B., LeCun, Y., and Culurciello, E.
(2012). NeuFlow: dataflow vision processing system-on-a-chip. In Circuits and Sys-
tems (MWSCAS), 2012 IEEE 55th International
Midwest Symposium on, pages
1044–1047. IEEE.