Probabilistic graphical models (PGMs) are a rich framework for encoding probability distributions over complex domains: joint (multivariate) distributions over large numbers of random variables that interact with each other. These representations sit at the intersection of statistics and computer science, relying on concepts from probability theory, graph algorithms, machine learning, and more. They are the basis for the state-of-the-art methods in a wide variety of applications, such as medical diagnosis, image understanding, speech recognition, natural language processing, and many, many more. They are also a foundational tool in formulating many machine learning problems.
About this Course
학습자 경력 결과
완료하는 데 약 30시간 필요
학습자 경력 결과
완료하는 데 약 30시간 필요
The Leland Stanford Junior University, commonly referred to as Stanford University or Stanford, is an American private research university located in Stanford, California on an 8,180-acre (3,310 ha) campus near Palo Alto, California, United States.
- 5 stars
- 4 stars
- 3 stars
- 2 stars
- 1 star
PROBABILISTIC GRAPHICAL MODELS 1: REPRESENTATION의 최상위 리뷰
Overall very good quality content. PAs are useful but some questions/tests leave too much to interpretation and can be frustrating for students. Audio quality for the classes could also be improved.
The lecture was a bit too compact and unsystematic. However, if you also do a lot of reading of the textbook, you can learn a lot. Besides, the Quiz and Programming task are of high qualities.
Prof. Koller did a great job communicating difficult material in an accessible manner. Thanks to her for starting Coursera and offering this advanced course so that we can all learn...Kudos!!
The course was deep, and well-taught. This is not a spoon-feeding course like some others. The only downside were some "mechanical" problems (e.g. code submission didn't work for me).
Excellent course, the effort of the instructor is well reflected in the content and the exercices. A must for every serious student on (decision theory or markov random fields tasks.
Superb exposition. Makes me want to continue learning till the very end of this course. Very intuitive explanations. Plan to complete all courses offered in this specialization.
Simply excellent. A wonderful course to begin the representation of PGM. Be advised.... this can get quite advanced. It's all about that Bayes, 'bout that Bayes.... no trouble.
I have Actually Earned Three Years of my life (at least) and one possible patent because of this course.\n\nThank You Daphne Mam. God Bless Everybody Associated with it.
learned a lot. lectures were easy to follow and the textbook was able to more fully explain things when I needed it. looking forward to the next course in the series.
Some parts are challenging enough in the PAs, if you are familiar with Matlab this course is a great opportunity to get familiar with PGMs and learn to handle these.
really great course! very clear and logical structure. I completed a graphical models course as part of my master's degree, and this really helped to consolidate it
Very well designed. There were areas here I struggled with the technical details and had to read up a lot to understand. The assignments are very well designed.
Probabilistic Graphical Models 전문 분야 정보
자주 묻는 질문
강의 및 과제를 언제 이용할 수 있게 되나요?
강좌에 등록하면 바로 모든 비디오, 테스트 및 프로그래밍 과제(해당하는 경우)에 접근할 수 있습니다. 상호 첨삭 과제는 이 세션이 시작된 경우에만 제출하고 검토할 수 있습니다. 강좌를 구매하지 않고 살펴보기만 하면 특정 과제에 접근하지 못할 수 있습니다.
이 전문 분야를 구독하면 무엇을 이용할 수 있나요?
강좌를 등록하면 전문 분야의 모든 강좌에 접근할 수 있고 강좌를 완료하면 수료증을 취득할 수 있습니다. 전자 수료증이 성취도 페이지에 추가되며 해당 페이지에서 수료증을 인쇄하거나 LinkedIn 프로필에 수료증을 추가할 수 있습니다. 강좌 내용만 읽고 살펴보려면 해당 강좌를 무료로 청강할 수 있습니다.
환불 규정은 어떻게 되나요?
재정 지원을 받을 수 있나요?
Learning Outcomes: By the end of this course, you will be able to
Apply the basic process of representing a scenario as a Bayesian network or a Markov network
Analyze the independence properties implied by a PGM, and determine whether they are a good match for your distribution
Decide which family of PGMs is more appropriate for your task
Utilize extra structure in the local distribution for a Bayesian network to allow for a more compact representation, including tree-structured CPDs, logistic CPDs, and linear Gaussian CPDs
Represent a Markov network in terms of features, via a log-linear model
Encode temporal models as a Hidden Markov Model (HMM) or as a Dynamic Bayesian Network (DBN)
Encode domains with repeating structure via a plate model
Represent a decision making problem as an influence diagram, and be able to use that model to compute optimal decision strategies and information gathering strategies
Honors track learners will be able to apply these ideas for complex, real-world problems
궁금한 점이 더 있으신가요? 학습자 도움말 센터를 방문해 보세요.