Lecture 1 - PyTorch Basics & Linear Regression

I also got the same error…now I am using my base conda enviroment…in that I am not getting any error

1 Like

@PrajwalPrashanth when I try to install the pip jovian it came out an error is “Some pip package fail to install” as the picture below… Is this causing the problem I mention above

When you say you’re running your base conda environment what do you mean? I’m getting the same error as alvertosk84

1 Like

https://jovian.ml/aakashns/machine-learning-intro This is the first notebook. Don’t worry you if you’re facing issues with installation. You have option to run the notebook on kaggle/colab.

I will update the list if installation instructions later on


Great Session Akash!! Thanks!!

1 Like

Do we get a confirmation that we have attended the first lecture?


Just finished the lecture 1, have been learning python for over 4 months and now i think data science is what I’m gonna pursue my career on. Thank you so much. Looking forward in practicing right now and learning more.


it was really good session to kick start for all !!! . Thanks Aakash !!!


How mark my

Hey, great lecture but it left me with a couple questions:

  1. What happens if the weight, by chance, starts at a local maxima? Since the derivative would be zero, which direction would the weight change?
  2. Can someone clarify what .backward() actually does?
  3. Is there any “rule of thumb” surrounding epoch amount, training set length, and learning size/step size?
  4. Why is squared loss used as opposed to the absolute value?

It was an amazing session sir .
thanks a lot for making it free.

  1. Then you subtract the zero from the weights, so they stays the same.
  2. It’s to actually calculate the gradients (I think they’re not calculated automatically when performing operations).
  3. Sometimes you train until your loss stops decreasing (with unspecified amount of epochs). What’s really important is the data - the more the better.
  4. Just a preference, it could be used as well (it’s called L1 loss or sth in PyTorch).
1 Like

Finally found the attendance button. lol

1 Like

Ah, thanks for the answers. If gradients aren’t automatically calculated, what does enable_grad do?
Is it just to “flag” them to be calculated when .backwards() is called?

I suppose you think about requires_grad. Yes, without it the .grad is None.

I also did get the same issue.

Was anyone able to solve this issue while running on conda?

where i can mark my attendance?

At the top of this page, there is an option to mark your attendance.

1 Like

yes, you can watch anytime