Free shipping on all orders over $50
7-15 days international
19 people viewing this product right now!
30-day free returns
Secure checkout
23655564
This book, a concise introduction to modern probability theory and certain of its ramifications, deals with a subject indispensable to natural scientists and mathematicians alike. Here the readers, with some knowledge of mathematics, will find an excellent treatment of the elements of probability together with numerous applications. Professor Y. A. Rozanov, an internationally known mathematician whose work in probability theory and stochastic processes has received wide acclaim, combines succinctness of style with a judicious selection of topics. His book is highly readable, fast-moving, and self-contained.The author begins with basic concepts and moves on to combination of events, dependent events and random variables. He then covers Bernoulli trials and the De Moivre-Laplace theorem, which involve three important probability distributions (binomial, Poisson, and normal or Gaussian). The last three chapters are devoted to limit theorems, a detailed treatment of Markov chains, continuous Markov processes. Also included are appendixes on information theory, game theory, branching processes, and problems of optimal control. Each of the eight chapters and four appendixes has been equipped with numerous relevant problems (150 of them), many with hints and answers. This volume is another in the popular series of fine translations from the Russian by Richard A. Silverman. Dr. Silverman, a former member of the Courant Institute of Mathematical Sciences of New York University and the Lincoln Laboratory of the Massachusetts Institute of Technology, is himself the author of numerous papers on applied probability theory. He has heavily revised the English edition and added new material. The clear exposition, the ample illustrations and problems, the cross-references, index, and bibliography make this book useful for self-study or the classroom.
"Concise" is indeed the operative word here. This book is probably not suitable as a first text on the subject, but makes an excellent review or quick reference for the topics it covers.Essentially, this text is geared toward taking someone who has - in principle - no knowledge of probability and introducing them specifically to Markov processes. There is very little attention paid to conditional probabilities, and Bayes' rule is never even mentioned.Also, this book requires no measure theory.Chapter 1 covers basic concepts: probability as relative frequency, sampling with and without replacement, binomial and multinomial coefficients.Chapter 2 is titled "Combination of Events". It introduces the idea of the sample space, and focuses on how probability interacts with set theoretic operations such as intersection and union. It ends with a proof of the First Borel-Cantelli Lemma.The third chapter introduces independence and ends with a proof of the Second Borel-Cantelli Lemma.The Borel-Cantelli Lemmas are somewhat technical results that are needed to the get the theory of Markov processes off the ground, so it's pretty clear where this book is headed early on. The proofs of both of the lemmas are very tidy.Chapter 4 is devoted to random variables. Here we find the definitions of expectation, variance, and the correlation coefficient along with Chebyshev's Inequality.Chapter 5 covers the Bernoulli distribution, the Poisson distribution, and the Normal distribution. We are also treated to the De Moivre-Laplace theorem as a stepping stone toward the Central Limit Theorem.Chapter 6 is titled "Some Limit Theorems". We are immediately provided with the proof and then statement - in that order - of the Weak Law of Large Numbers. We are then provided merely with the statement of the Strong Law of Large Numbers. This chapter then introduces Generating Functions which are used quite heavily in the remainder of the work. This chapter also introduces Characteristic Functions, which don't get much attention and concludes with the Central Limit Theorem.Chapter 7 introduces Markov Chains while chapter 8 covers Continuous Markov Processes and naturally covers the Chapman-Kolmogorov equations. Here simply called the Kolmogorov equations for the fairly obvious reason that the author is Russian.The book ends with four short appendices which introduce the reader in turn to the following topics: Information Theory, Game Theory, Branching Processes, and Optimal Control. I thought these were wonderful although obviously none of them covers very much ground.This book is actually quite delightful especially for someone who already has some background in basic probability. It does provide and good and very quick introduction to Markov processes, but it's scope of coverage of any topic is necessarily quite limited.