3 déc. 2018
Extremely helpful review of the basics, rooted in mathematics, but not overly cumbersome. Very clear, and example coding exercises greatly improved my understanding of the importance of vectorization.
6 mars 2019
I understand all those thing which you have discussed in this course and I also like the way first tell story of concet and assign assignment. Now I fall in love with neural network and deep learning.
par Baili O•
23 janv. 2018
This is a great course which covers some popular machine learning techniques such as regression. And then it moves to deep learning with neural networks with the techniques of forward and backward propagation. It is a good course for beginner and the homework are fairly easy. However, this course still leaves some unanswered questions which might be covered in the future courses such as how to select hyperparameters, why we choose this specific cost function, are there any other deep learning framework other than neural networks structure, any other application other than image recognition. In addition, for those who have some background in machine learning, the interview section is a bonus which talked about GANs.
There are somethings not very enjoyable as well. For example, the notebook is unable to download so people have to write it down otherwise when the course expired, it is quite hard to get the course material. The lecture notes are badly organized: you have to download every slides one by one. (or I just didn't find the right place to download). Thirdly, this course didn't talk too much about techniques that the industry is using. What I am trying to say is that I don't know if the techniques in this course is applicable in the industry, is it too simple or is it too old etc.
Overall, it is a very fun and educated course. I can't wait to jump into the next course.
par Melissa B•
16 août 2020
This course has been very interesting and engaging for me. Dr. Ng explains everything very thoroughly and provides compelling examples of real world application for the material. Occasionally he can be a bit redundant, but I found that helpful, since sometimes it takes more than one pass through the material to understand it clearly. I am also taking his Machine Learning course (Stanford) concurrently and I found the courses to complement each other very well.
Additional practical notes: The linear algebra involved in the course is relatively basic and explained thoroughly enough that it can be picked up along the way if you don't have much math background. Also, the programming assignments help expose you to Python syntax. If you don't have much Python experience, use the discussion forums. If you do, you'll find the assignments to be incredibly easy, as they have ample starter code with a few plug-and-chug "your code here" sections. These are mostly just to demonstrate how the material is applied. The primary value in the course is conceptual. For someone with very little coding experience like myself, I appreciate how thoroughly the code was annotated so I could get a grasp on what it was doing without having to stab through trying to produce it myself.
Overall, I really loved this course, and I learned so much!
par Felipe P C N•
10 juil. 2020
O curso concilia uma abordagem compreensiva - partindo de conceitos intuitivos/elementares - com uma profundidade respeitável - construindo com cuidado técnicas mais aprofundadas. As aulas são extremamente didáticas e bem elaboradas, o que torna a familiarização com os conceitos de redes neurais (e a matemática por trás deles) mais natural. As avaliações têm tanto um componente teórico (quizzes, mais voltados para verificação de conhecimentos) quanto um prático (exercícios de programação). Por mais que seja positivo haver o componente prático, eu diria que ele poderia ser um pouco menos "guiado" - na prática, o que eles pedem é que o aluno "complete as lacunas" em um programa já (muito bem) estruturado (e praticamente completado) pelos professores. Senti a necessidade de complementar esses exercícios com tentativas por conta própria, programando os algoritmos do zero, para de fato assegurar que consigo implementar o que aprendi. Esse passo é sensivelmente mais difícil que os programas do curso (afinal, implementar um projeto do zero é essencialmente desafiador). Mesmo assim, vejo que esse tipo de "iniciativa" de levar os conteúdos aprendidos a alguma aplicação/prática para além do pedido pelo curso é algo que idealmente deveria ser feito ao se estudar qualquer disciplina.
28 févr. 2020
This course from the basic to the advanced, leading us to understand deep learning. From the initial logistic regression, then to the shallow neural network, and finally to the deep neural network, we gradually learned the neural network representation and calculation process, and finally began to implement the cat image recognition binary classifier. The course is very clear and logical, eliminating the tedious mathematical derivation, but still allowing us to understand all the mathematical details including calculation and vectorization. The assignments are done step by step, starting from the basic functions and gradually encapsulating, and finally constitute a complete neural network, which enables students to have a deep understanding of neural network and master knowledge from practice. It is worth mentioning that the setting of course gradient is reasonable, and the details that are difficult to understand in the previous course do not need to be understood all at once. In the later courses, the understanding of the previous knowledge points will be deepened repeatedly, and due to the foreshadation of other knowledge points, a more complete and comprehensive supplement will be provided to the previous knowledge. Looking forward to series two.
par Ronald A R•
17 déc. 2021
It would be difficult to overstate the value of truly laboring through this very well designed and conducted course. I've an advanced degree in cell/molecular biology with thesis work in immunochemistry and physical biochemistry (UCSD). I did undergraduate work including the regular chemistry coursework including thermodynamics, statistical mechanics, & quantum mechanics. I consider myself fairly mature at applied math in my areas of training and work. But while not ill prepared, it took some real effort to gain a deeper understanding of the application of the underlying math for backward propagation. This course with a little help from 3brown1blue, provided a great introduction to the "mysteries" of NN, "deep learning", and AI. Dr. Ng provides an most pleasant presence as he works to give a new student access to what is arguably among the most sophisticated and promising realms in computer and artificial cognitive science. I could not have gained access to this space without his course. I will move on to the next course in the sequence of four. BTW, I'm 75 years old, semi-retired, and pursuing this only out of the desire to know and understand. Thanks yet again to Dr. Ng, his colleagues (named and unnamed), and Coursera.
par Xiao G•
7 oct. 2017
This is my first completed course on Coursera.org!!! and I win a certificate with nearly full marks! I was very bad at coding .... although now I still not good at it, this course convinces me that even a poor guy in coding like me can finish and establish a neural networks! it truly boosts my confidence! Thanks Andrew Ng.
One thing to note for this course.... maybe it can improve later in this side. I feel quite easy and comfortable with all mathematical deductions.... however, when I code in week 4, the backward propagation, I nearly get lost.... I spend the whole afternoon and night to solve it and then finish assignment 2 in week 4 (from 2 pm till 10:03 pm)... I bet many people get stuck on that area. hmm, It's really easy to get lost and puzzled there. I guess there maybe some points we can figure it out more easily. Just a little advice.
Anyway, I love this course. It's a trigger for my coding area, though I'm a physics student and now in electrical engineering.... I still feel very comfortable about this course.
Thanks Andrew Ng!!! I don't know how to express my gratitude to you! If I did not take this course, I may never ever to attempt on Deep Learning, such a complex and advanced thing.
I will continue to study~~ see you
par Luca C•
26 janv. 2019
Pros: + You will understand clearly how things work and why they work
+ Provide mathematical insights for those who are interested (really a big plus wrt other courses)
+ Overall a simple introduction to Neural Nets. Even those who already have experience can benefit from this brush up (even at a fast pace: you can complete it in 2-4 days)
Cons: - Since it is quite basic material, those who are already accustomed to NN might want to jump to the second course of the specialization.
- You learn python by doing, but you will not get an understanding of python. I would suggest to get a little bit familiar with it in other ways (however, this is not a requirement to master the course).
Clear, quick overview of the basics of Neural Network. Provide even some mathematical justification, even if it is really not a requirment to fully understand it in order to succesfully achieve the course. However, IMHO it is always good to have at least some insights of the mathematics behind.
This course sets well the basics, but to be really able to work on your own projects I think it is a must to take the second course of this specialization.
par Amit R B•
27 nov. 2019
This course is truly deserving of its high ratings. Prof. Andrew Ng's extensive breakdown of the structure and function of neural networks work is unparalleled. For me personally this course has been of great help. The theory lectures made me understand just how these networks "learn". This course is a great beginning and I think, prepares the student well to learn more in depth and advance concepts of deep learning.
However, if you are looking to get hands-on experience building and training deep learning modes I would recommend checking out some free resources on YouTube with the Keras framework. I played around with with Keras following the YouTube channel Sentdex's Keras tutorials. Then took this course to get a more mathematical and theoretical understanding. Some students might find themselves a bit unprepared for the coding exercises, since the lectures are more focused on theory and math, showing little to no code. This is why I thinks this is a great (if not the best) 2nd course, but maybe not as helpful as a first introduction.
For a free first introduction, check out the channel 3blue1brown's videos on Neural Networks to get your feet wet, before diving further deep. ;P
8 juin 2020
I am seeing that people are complaining in the top review saying that the course is bad because there isn't stoppages for you to think and all that. Or like the code is relatively simple to write.
What I want to say is that this is how, in general, what you learn in school. And this is the exact style that is being taught in school. The professor isn't going to pause here and there for you, unless you raise your hand to ask questions. In my machine learning class there isn't even quiz to check your understanding. The only thing missing in this course is PROJECT. Which is something that a online course can't provide, and there is no way the tutors can mark all your project.
When you complain about the lab assignments, this is exactly the type of lab assignment that you get in school, just that perhaps in school the variables isn't even given (but that is because tutors are within range and you can ask them any time you want, but this isn't the case for online learning, they have to make this do-able for everyone)
For something that is free, I can't believe that people are complaining about the quality, this is on par with my school's standard, and my school isn't bad at all.
par Randall S•
5 oct. 2017
Dr. Andrew Ng is brilliant and it is so amazing to have access to this type of knowledge for less than I spend on Starbucks in two to three weeks. I am taking some online courses at a big name university (to the tune of $4,000 per course), and for the money, this is a real bargain and just as good if not better!
The thing I liked most about this particular course is that it showed us what's happening under the hood, and not just a course on how to use tools, nor is it all theory. Dr. Ng also introduced us to Geoffrey Hinton, the pioneer of backward propagation which was worth the price of admission alone.
That said, it was not so tough that I couldn't keep up. I would say that having some exposure to calculus would help, but it is not required. Also, you need to be more than just familiar with Python, but if you can spend a few extra hours per week on the course, you can work your way through it with just a familiarity with Python.
It has challenged me to keep going to the next level and complete the specialization -- AI is not rocket science -- at least not at the level of applying this knowledge. Being at Dr. Ng's level might be a different story.
par Anand R•
29 janv. 2018
To set the context, I have a PhD in Computer Engineering from the University of Texas at Austin. I am a working professional (13+ years), but just getting into the field of ML and AI.
I completed Dr. Ng's course on Machine Learning on Coursera first. I recommend that students of this course should first complete that course (or an equivalent one). This course was an excellent review of the basic concepts of Neural Networks. The lectures were well presented and the maths/equations were explained intuitively. The problem solving assignments were in Python (as opposed to Matlab). As before, Dr. Ng walked us through the assignments: hand-holding us through the solution. The quizzes were fairly challenging and helped me reinforce the concepts quite well.
I wish there were a few open problems (Kaggle style) at the end of the course so that the class students could compete with each other. It would be a good addition to the course. I would appreciate more real world examples throughout the course as well.
I look forward to completing the remaining courses! Thank you, Dr. Ng. Thanks you, teaching assistants. Thank you, Coursera. This is truly a wonderful course.
par MD A•
18 juil. 2019
Thorough and simple explanations that help internalize the deep learning concepts. Video lectures are very helpful. Listen more than once to clarify concepts. Very useful jupyter notebook exercises with solutions that provide knowledge reinforcement. Vectorized form of deep learning neural network equations enable development of clutter-free and faster scalable solutions. Before taking the course refresh your knowledge of linear algebra esp. basic matrix operation such as matrix size, transpose, and implementation in Python via numpy such as numpy.dot for matrix multiplication, numpy.multiply for element-wise multiplication. Familiarity of Python key:value dictionary data structure and retrieval of values via keys. This knowledge will build confidence to code the functions and methods for forward propagation, back propagation, and gradient descent to update weights and biases. Also pay some attention to how indices in square brackets are used to identify matrices for inputs, outputs, parameters (weights and biases), activation values/models, various layers of a neural network, and nodes in a particular layer (all explained well in lectures.
par Alex M•
15 oct. 2018
I've been impossibly busy and first thought this was something i could play in the background while I did other work. Quickly it became apparent that data I had been used to with M.shape = (user/observation/etc, feature) was now the transpose. This took a simple few examples on paper to convince me why this was a superior notation for D/RNN architectures given numpy notation. I also at first thought that the bias should be added to W, X, for greater expressibility of the relationship y = g(WX) and for the backprop updates that require 'estimating' the W.T*g^-1(y) and g^-1(y)*X.T (where y is understood as the general activation after layer l and X is the general output of the previous layer), but now I see why separating the bias is useful -- it estimates the 'scale' of all the data at the output layer at once (estimating the unbalance in the marginal distribution, for example), whereas the other gradients come from estimating the perturbative deformation in the input layer, thus they are slightly different from the perspective of forward backwards distributional learning. Bravo, and thank you!
par Mahesh G•
29 août 2017
Thanks for the course. Very neatly explained on the background maths that happens in neural networks. This course will help you understand the step by step what happens within the network. The step by step procedure which is explained by Professor is great and he has repeatedly stressed the important steps to make it clear. Along with the explaining the formulas the assignment helps in implementing the formulas step by step and converting the whole thing to a neural network model, this is a great learning. One of the important thing covered in the beginning of the course is about vectorization, python broadcasting which is the key for neural network.
The pace at which Professor explained the concepts is good and easy to follow and the structure of the course is well laid-out which helps for the beginners.
One thing that could have been better is the assignments, current assignments are definitely helpful for beginners like me, but could have some more assignments which increases the complexity level (may be it is there in subsequent courses).
Overall very good course and helped me
par Krishna k N•
18 mai 2019
I admire Professor Andrew Ng's patience in helping the students take baby steps by painting a big picture from each small pixel, just as how a neural network is built.
This course has given me great exposure to how neural network, although I realize I need to take a Python course to type code more freely and easily.
I'm going to do that next and then come back to the remaining courses in this specialization.
feedback - it's really hard to visualize some of these matrices and their dimensions used in a large neural network with so many parameters such as nx features, m training examples, n iterations, L layers with (nL, NL-1) weights, (nL,1) biases etc. I understand it's hard to show these matrices by writing as they are very large. I wish someone would develop a more "animative" way of illustrating these matrices that will make the intuition more stronger. for example, calculating forward_activation for all layers and all neurons across these layers by just passing X and parameters is a massive operation and the intuition stumbles purely by the scale of such a matrix operation.
par Eung P•
9 mars 2021
This is a really excellent course. I am surprised to have learned the fundamentals of deep learning in only one month! I am very happy that I could try myself a real-world problem on my PC w/o GPU. This was totally beyond my expectations.
Prof. Ng teaches all the details step-by-step without any logical jump. Even those who does not know high-level math can catch up. He tried to relieve students' concerns and burdens, any time I lost confidence. After completing the Programming Assignments, I could be comfortable. The assignments come with detailed explanations so that students can follow up without much trouble.
His fast hand-writing was sometimes a little hard to read (e.g. "[", ), but typed summaries could clear them out.
Having an interview with Prof. Jeff Hinton was also another big addition to this course. Actually, this was the time I started to gain more trust in this course. Frankly I didn't know about Coursera or Prof. Ng before.
The mentors (esp. Paul) are excellent with quick responses with sufficiently detailed guides.
Without any hesitation, I throw 5+ stars.
par Nikos S•
23 déc. 2019
I am a CS student with an interest in AI; therefore, I was no stranger to Neural Networks even before I finished the first course. I decided to choose the DL specialization (my first in coursera) to learn more about NN and perhaps also to solve some questions I had (e.g. "Why a NN with an input layer, a hidden layer and an output layer is called 2-layer NN?" ). Proffessor Andrew Ng answered those questions and also made me think of the NN more intuitively rather than mathematically.
I found really impressive that while students with no prior knowledge in calculus can take this course with no problem, others, with good knowledge of calculus, like myself, are not bored even when the most basic concepts are explained (such as derivatives). Also, I liked a lot the cat image binary classification exercises, especially since the classifier was able to classify the well-know cat meme image as a cat picture :P .
I am looking forward for the next courses of the specialization to also learn about other DL concepts, such as momentum and regularization, I am not already familiar with.
par ANUJ K J•
27 janv. 2020
Things that I learned: 1) Introduction to deep learning 2) Logistic Regression, gradient descent on logistic regression 3) Forward and Backward Propagation 4) Computational graph and how to use the computational graph to calculate the forward and backward prop 5) Shallow Neural Network, and how to work on it end to end 6) Deep Neural Network and it's end-to-end implementation on an application (Classification of cat vs non-cat). Pros: 1) The course is in Python 2) The way Andrew Ng sir teaches is simplistic and memorable as he starts from small concepts and relates the same concept in complex problems too. For example:- He started teaching forward and backward prop using logistic regression then he carried the same in shallow neural net and finally on a deep neural net. Cons: 1) The number of questions in the video is less as compared to the Machine Learning course by Andrew Ng. 2) There could have been more clarification on a few small topics. 3) The video didn't provide any sort of notes or lecture slides. ( Note that I found useful: https://lnkd.in/f3fZGCy )
par Peter D•
3 déc. 2017
As usual, Prof. Andrew Ng knocks it out of the park!!! He would argue otherwise, but he's a natural born teacher whether he admits it or not. This was a challenging course, but I found the objectives to be achievable with a bit of hard work and cool-headed thought. Having taken Prof. Ng's Machine Learning course already, most of the material from the first two weeks of NN4DL was review. Unlike the broader ML course, DL was much more narrowly focused on concepts leading to mastery of deep neural nets. It also ditches the MATLAB/Octave used in ML for a more portable Python environment. I had basically no knowledge of Python when I started, so I guess I learned it in 4 weeks! :D My advice: take ML first, or you may be lost. I had the math and ML background for this stuff to make sense, so Python was the only thing entirely new to me. If you're fuzzy on calculus, or ML, or programming, I don't recommend starting with this course. But if you have a strong background on those things, you'll find this course is well worth your time! Good luck.
par Rob M•
12 mai 2019
I've taken and finished Udacity's Nanodegree, and while it certainly has a lot of its own strengths, I came here to get another perspective on the math involved, especially in backpropogation and numpy operations. Lo and behold, this class (Andrew in particular, of course) delivered exactly what I was looking for. And because the course was supremely self-paced, instead of feeling rushed to hit an official deadline like Udacity's course, I was able to take the time I needed to watch the videos a couple times each, when necessary, and really drill home the concepts.
Lastly, the projects here at Coursera are extremely well thought out, organized, and testable. I *loved* the use of the numpy seed operation, so when I completed a function and tested it, I felt extremely confident that the inputs, operations, and outputs were exactly what I needed. At this point, I definitely like the approach to projects much better than Udacity's (always felt like more of a guessing game there).
I'm excited to start and finish the next course in the Specialization!
par Michael S E•
28 févr. 2018
Excellent course. Quick introduction to the basics of neural networks. This course has very high overlap with Prof. Ng's course on neural networks at Stanford. This appears to be the updated version on his new DeepLearning.ai platform.
The programming assignments are very user friendly, in that the code is already highly structured with student code just to fill in a few blanks. They also provide built-in test cases. The difficulty level is not high compared to a more open ended problem formulation (let alone a real world task). The assignments do make efficient use of student time in that they focus on the essential aspects of the course material and minimize time spent on extraneous computer programming challenges.
I appreciated the consistent and strategically chosen notation, which makes it easier to translate formulas into code snippets. Ng's notation conventions allow you to make an educated guess at how to vectorize algorithms in numpy simply by capitalizing variable names.
Thank you for sharing your knowledge and expertise with us!
par Manuel G•
9 sept. 2017
This is a great class to get introduced to deep learning concepts and get some hands on experience with the underlying machine learning aspects. The Jupyter notebooks are great in that you are left with something you can use later as a starting point if you want to do your own implementations. The flip side of that is that, in my view, the coding assignments are made too easy and I feel that after all the hints and given the code you are given, the student's contribution is a tad too trivial at that point. Still, this doesn't change my rating because from the perspective of learning about DL concepts, this is not a crucial point. Since the course is still very new, there remain a few bits of consistency in notation and other little details that haven't yet been 100% fixed, but there's a lot of activity in the forums to help you clarify things and give feedback on what is not working.
As usual, Andrew Ng does a great job of motivating and explaining all the concepts. If you enjoyed his ML class, definitely go with this specialization.
par Tony H•
16 août 2017
Extremely well-taught and well-structured introduction to neural networks and deep learning. I found the explanations of forward and back propagation to be at a level suitable for getting the algorithms to work without swamping one in detailed calculus, but with enough detail to enable productive further study. There is an introduction to computation graphs that will hopefully lead into Tensorflow in the next courses in this specialisation. Professor Ng is a methodical, very knowledgable and interesting teacher and I really enjoyed all his video lectures. The weekly quizzes are reasonably challenging and the programming exercises very well written and enjoyable. If I have one minor criticism it is that there is perhaps a little too much 'hand-holding' in the programming exercises; I felt that some code was supplied that could have been left for the student to fill in, some very basic Python instructions could also have been left for the student. I am greatly looking forward to the next courses in this specialisation.
par Shakti A P•
15 sept. 2022
This is an amazing course which has helped me learn so many things about Machine Learning and Deep Learning that I could've never imagined learning/trying to learn on my own. Andrew Ng is a brilliantly awesome teacher and an amazing person! The way he teaches tough concepts so easily is beyond my comprehension, but thanks to him, I now have so much of a better understanding and intuition about these in-depth concepts of Neural Networks and Deep Learning. I've had a blast completing this course and its assignments. Each lecture was like a fun learning experience I kept looking forward to. I'll try my very best to leverage my learnings from this course everywhere! I'll be sure to complete the other courses in the Deep Learning Specialization series. Taking up this course is probably the best academic choice I've made in quite a long time - it has kindled my interest in AI and Deep Learning like never before! Thanks so very much ~ Andrew, Coursera and all the amazing people behind this wonderful course!
par Nicolas G•
22 juil. 2020
The best course in deep learning out there! Python assignments might be a little easy for someone with extensive knowledge of python. I don't think this is necessarily a bad thing, because it makes the student focus more on understanding deep learning rather than spending time writing python-based code (e.g. creating the functions, importing libraries, etc). If you don't have extensive python knowledge, you will be able to do the assignments, but I would recommend to try to understand how the whole code is structured and what each line of code is doing. Also, although Andrew continuously suggests to not worry about the calculus, I think it is helpful (and maybe necessary) to have some basic knowledge of calculus (derivatives) and linear algebra (vectors, matrix multiplication, etc) before doing the course. This will help you have a deeper understanding of the math behind the code and how neural networks work. Overall, Andrew is a great professor and covers all topics in a comprehensive and understandable way!