CS
Structure and Interpretation of Computer Programs - 2nd Edition 谷歌图书 豆瓣 Goodreads
作者: Harold Abelson / Gerald Jay Sussman The MIT Press 1996 - 9
Structure and Interpretation of Computer Programs has had a dramatic impact on computer science curricula over the past decade. This long-awaited revision contains changes throughout the text.
There are new implementations of most of the major programming systems in the book, including the interpreters and compilers, and the authors have incorporated many small changes that reflect their experience teaching the course at MIT since the first edition was published.
A new theme has been introduced that emphasizes the central role played by different approaches to dealing with time in computational models: objects with state, concurrent programming, functional programming and lazy evaluation, and nondeterministic programming. There are new example sections on higher-order procedures in graphics and on applications of stream processing in numerical programming, and many new exercises.
In addition, all the programs have been reworked to run in any Scheme implementation that adheres to the IEEE standard.
Supervised Sequence Labelling with Recurrent Neural Networks 豆瓣
作者: Graves, Alex Springer 2012 - 2
Supervised sequence labelling is a vital area of machine learning, encompassing tasks such as speech, handwriting and gesture recognition, protein secondary structure prediction and part-of-speech tagging. Recurrent neural networks are powerful sequence learning tools―robust to input noise and distortion, able to exploit long-range contextual information―that would seem ideally suited to such problems. However their role in large-scale sequence labelling systems has so far been auxiliary.
The goal of this book is a complete framework for classifying and transcribing sequential data with recurrent neural networks only. Three main innovations are introduced in order to realise this goal. Firstly, the connectionist temporal classification output layer allows the framework to be trained with unsegmented target sequences, such as phoneme-level speech transcriptions; this is in contrast to previous connectionist approaches, which were dependent on error-prone prior segmentation. Secondly, multidimensional recurrent neural networks extend the framework in a natural way to data with more than one spatio-temporal dimension, such as images and videos. Thirdly, the use of hierarchical subsampling makes it feasible to apply the framework to very large or high resolution sequences, such as raw audio or video.
Experimental validation is provided by state-of-the-art results in speech and handwriting recognition.
Learning Deep Architectures for AI 豆瓣
作者: Yoshua Bengio
Theoretical results suggest that in order to learn the kind of complicated
functions that can represent high-level abstractions (e.g., in
vision, language, and other AI-level tasks), one may need deep architectures.
Deep architectures are composed of multiple levels of non-linear
operations, such as in neural nets with many hidden layers or in complicated
propositional formulae re-using many sub-formulae. Searching
the parameter space of deep architectures is a difficult task, but learning
algorithms such as those for Deep Belief Networks have recently been
proposed to tackle this problem with notable success, beating the stateof-
the-art in certain areas. This monograph discusses the motivations
and principles regarding learning algorithms for deep architectures, in
particular those exploiting as building blocks unsupervised learning of
single-layer models such as Restricted Boltzmann Machines, used to
construct deeper models such as Deep Belief Networks.
A Calculus of Ideas 豆瓣
作者: Ulf Grenander World Scientific Publishing Company 2012 - 9
This monograph reports a thought experiment with a mathematical structure intended to illustrate the workings of a mind. It presents a mathematical theory of human thought based on pattern theory with a graph-based approach to thinking. The method illustrated and produced by extensive computer simulations is related to neural networks. Based mainly on introspection, it is speculative rather than empirical such that it differs radically in attitude from the conventional wisdom of current cognitive science.
Convex Optimization 豆瓣 Goodreads
Convex Optimization
8.0 (5 个评分) 作者: Stephen Boyd / Lieven Vandenberghe Cambridge University Press 2004 - 3
Convex optimization problems arise frequently in many different fields. This book provides a comprehensive introduction to the subject, and shows in detail how such problems can be solved numerically with great efficiency. The book begins with the basic elements of convex sets and functions, and then describes various classes of convex optimization problems. Duality and approximation techniques are then covered, as are statistical estimation techniques. Various geometrical problems are then presented, and there is detailed discussion of unconstrained and constrained minimization problems, and interior-point methods. The focus of the book is on recognizing convex optimization problems and then finding the most appropriate technique for solving them. It contains many worked examples and homework exercises and will appeal to students, researchers and practitioners in fields such as engineering, computer science, mathematics, statistics, finance and economics.
Applied Optimal Estimation 豆瓣
作者: A Gelb MIT Press 1974 - 1
This is the first book on the optimal estimation that places its major emphasis on practical applications, treating the subject more from an engineering than a mathematical orientation. Even so, theoretical and mathematical concepts are introduced and developed sufficiently to make the book a self-contained source of instruction for readers without prior knowledge of the basic principles of the field. The work is the product of the technical staff of the The Analytic Sciences Corporation (TASC), an organization whose success has resulted largely from its applications of optimal estimation techniques to a wide variety of real situations involving large-scale systemsArthur Gelb writes in the Foreword that "It is our intent throughout to provide a simple and interesting picture of the central issues underlying modern estimation theory and practice. Heuristic, rather than theoretically elegant, arguments are used extensively, with emphasis on physical insights and key questions of practical importance."Numerous illustrative examples, many based on actual applications, have been interspersed throughout the text to lead the student to a concrete understanding of the theoretical material. The inclusion of problems with "built-in" answers at the end of each of the nine chapters further enhances the self-study potential of the text.After a brief historical prelude, the book introduces the mathematics underlying random process theory and state-space characterization of linear dynamic systems. The theory and practice of optimal estimation is them presented, including filtering, smoothing, and prediction. Both linear and non-linear systems, and continuous- and discrete-time cases, are covered in considerable detail. New results are described concerning the application of covariance analysis to non-linear systems and the connection between observers and optimal estimators. The final chapters treat such practical and often pivotal issues as suboptimal structure, and computer loading considerations.This book is an outgrowth of a course given by TASC at a number of US Government facilities. Virtually all of the members of the TASC technical staff have, at one time and in one way or another, contributed to the material contained in the work
Introduction to Semi-Supervised Learning 豆瓣
作者: Xiaojin Zhu / Andrew B. Goldberg Morgan and Claypool Publishers 2009 - 6
Semi-supervised learning is a learning paradigm concerned with the study of how computers and natural systems such as humans learn in the presence of both labeled and unlabeled data. Traditionally, learning has been studied either in the unsupervised paradigm (e.g., clustering, outlier detection) where all the data are unlabeled, or in the supervised paradigm (e.g., classification, regression) where all the data are labeled. The goal of semi-supervised learning is to understand how combining labeled and unlabeled data may change the learning behavior, and design algorithms that take advantage of such a combination. Semi-supervised learning is of great interest in machine learning and data mining because it can use readily available unlabeled data to improve supervised learning tasks when the labeled data are scarce or expensive. Semi-supervised learning also shows potential as a quantitative tool to understand human category learning, where most of the input is self-evidently unlabeled. In this introductory book, we present some popular semi-supervised learning models, including self-training, mixture models, co-training and multiview learning, graph-based methods, and semi-supervised support vector machines. For each model, we discuss its basic mathematical formulation. The success of semi-supervised learning depends critically on some underlying assumptions. We emphasize the assumptions made by each model and give counterexamples when appropriate to demonstrate the limitations of the different models. In addition, we discuss semi-supervised learning for cognitive psychology. Finally, we give a computational learning theoretic perspective on semi-supervised learning, and we conclude the book with a brief discussion of open questions in the field. Table of Contents: Introduction to Statistical Machine Learning / Overview of Semi-Supervised Learning / Mixture Models and EM / Co-Training / Graph-Based Semi-Supervised Learning / Semi-Supervised Support Vector Machines / Human Semi-Supervised Learning / Theory and Outlook
An Introduction to Kolmogorov Complexity and Its Applications 豆瓣
作者: Ming Li / Paul M.B. Vitányi Springer 2008 - 11
'The book is outstanding and admirable in many respects ...is necessary reading for all kinds of readers from undergraduate students to top authorities in the field' - "Journal of Symbolic Logic". Written by two experts in the field, this is the only comprehensive and unified treatment of the central ideas and applications of Kolmogorov complexity. The book presents a thorough treatment of the subject with a wide range of illustrative applications. Such applications include the randomness of finite objects or infinite sequences, Martin-Loef tests for randomness, information theory, computational learning theory, the complexity of algorithms, and the thermodynamics of computing. It will be ideal for advanced undergraduate students, graduate students, and researchers in computer science, mathematics, cognitive sciences, philosophy, artificial intelligence, statistics, and physics. The book is self-contained in that it contains the basic requirements from mathematics and computer science. Included are also numerous problem sets, comments, source references, and hints to solutions of problems. New topics in this edition include Omega numbers, Kolmogorov-Loveland randomness, universal learning, communication complexity, Kolmogorov's random graphs, time-limited universal distribution, Shannon information and others.
Probably Approximately Correct 豆瓣 Goodreads
作者: Leslie Valiant Basic Books 2013 - 6
How does life prosper in a complex and erratic world? While we know that nature follows patterns - such as the law of gravity - our everyday lives are beyond what known science can predict. We nevertheless muddle through even in the absence of theories of how to act. But how do we do it? In Probably Approximately Correct, computer scientist Leslie Valiant presents a masterful synthesis of learning and evolution to show how both individually and collectively we not only survive, but prosper in a world as complex as our own. The key is "probably approximately correct" algorithms, a concept Valiant developed to explain how effective behavior can be learned. The model shows that pragmatically coping with a problem can provide a satisfactory solution in the absence of any theory of the problem. After all, finding a mate does not require a theory of mating. Valiant's theory reveals the shared computational nature of evolution and learning, and sheds light on perennial questions such as nature versus nurture and the limits of artificial intelligence.
Parallel Distributed Processing - 2 Vol. Set 豆瓣
作者: David E. Rumelhart / James L. McClelland The MIT Press 1987 - 7
What makes people smarter than computers? These volumes by a pioneering neurocomputing group suggest that the answer lies in the massively parallel architecture of the human mind. They describe a new theory of cognition called connectionism that is challenging the idea of symbolic computation that has traditionally been at the center of debate in theoretical discussions about the mind. The authors' theory assumes the mind is composed of a great number of elementary units connected in a neural network. Mental processes are interactions between these units which excite and inhibit each other in parallel rather than sequential operations. In this context, knowledge can no longer be thought of as stored in localized structures; instead, it consists of the connections between pairs of units that are distributed throughout the network. Volume 1 lays the foundations of this exciting theory of parallel distributed processing, while Volume 2 applies it to a number of specific issues in cognitive science and neuroscience, with chapters describing models of aspects of perception, memory, language, and thought. David E. Rumelhart is Professor of Psychology at the University of California, San Diego. James L. McClelland is Professor of Psychology at Carnegie-Mellon University. A Bradford Book.