Welcome

I’m a .NET Software Developer in EuroForm A/S and have been involved in projects based on

  • ASP.NET Web Forms
  • ASP.NET Web API
  • ASP.NET MVC
  • WPF
  • others, like basic web tech (Html, Css, Js), Sql server, etc.

Recently, I’ve gain interest in learning, reading, practicing about Deep Learning, Computer Vision and hence, this website to share what I learnt and worked on.

I’ve taken couple of online courses to gain knowledge and grasp concepts about Deep Learning, Computer Vision and what I’ve learnt from them helps me in the thought process about a problem statement on table.

The goal for this website is to write blogs and share projects with others who are interested in this field.


Recent Posts

Time-series forecasting

20 Jul 2020
machine-learning forecasting

Forecasting

It is the technique to predict future data analysing historical data.

Type of forecasting

Quantitative forecasting Qualitative forecasting
Based on the data and any repeating historical patterns in the data. This procedure is taken when data is not available and historical patterns do not repeat. This is based on expert decision.
Captures complex patterns which humans cannot identify. May not capture complex patterns.
No bias Bias
Eg. Time-series forecasting Eg. Delphi method
More...

General ML questions

09 Jul 2020
machine-learning practice-questions

What is Linear regression?

  • Linear regression is the machine learning(ML) algorithm which finds the best linear-fit relationship on any given data, between independent and dependent variables.
  • It is mostly done by the Sum of Squared Residuals method.
  • Sum of the residuals of a linear regression model is always zero.

What are the assumptions in a Linear regression model?

  1. Assumption about the form of the model: Linearity assumption: It is assumed that there is a linear relationship between the dependent and independent variables.
  2. Assumptions about the residuals: a. Normality assumption: It is assumed that the error terms, are normally distributed. b. Zero mean assumption: It is assumed that the residuals have a mean value of zero. c. Constant variance assumption:
    • It is assumed that the residuals have same (but unknown) variance i.e. .
    • This assumption is also known as the assumption of homogenity or homoscedasticity. d. Independent error assumption: It is assumed that the residuals are independent of each other i.e. their pair-wise covariance is zero.
  3. Assumptions about the estimators: a. The independent variables are measured without error. b. The independent variables are independent of each other i.e. there is no multi-collinearity in the data.

What is Feature engineering? How do you apply it in the process of modelling?

  • Feature engineering is the process of transforming raw data into features that better represent the underlying problem to the predictive models, resulting in improved model accuracy on unseen data.
  • In simple words, it is the development of new features that may help you to understand and model the problem in a better way.
  • Types of feature engineering:
    • Business-driven:
      • This revolves around the inclusion of features from a business point of view.
      • The job here is to transform the business variables into features of the problem.
    • Data-driven:
      • In this, the features you add do not have a significant physical interpretation, but they help the model in prediction of the target variable.
  • To apply feature engineering, one must be fully acquainted with the dataset.
    • This involves knowing what the given data is, what it signifies, what the raw features are, etc.
    • You must also have a crystal clear idea of the problem, such as what factors affect the target variable, what the physical interpretation of the variable is, etc.

What is the use of regularisation? Explain L1 and L2 regularisation techniques?

  • Regularisation is the technique that is used to tackle the problem of overfitting of the model.
  • At times, the simple model might not be able to generalize the data and the complex model overfits. To address this problem, regularisation is used.
  • Regularisation is nothing but adding the coefficient terms (s) to the cost function so that the terms are penalised and are small in magnitude.
  • This essentially helps in capturing the trends in the data and at the same time prevents overfitting by not letting the model become too complex.
  • L1 or LASSO regularisation:
    • Here, the absolute values of the coefficients are added to the cost function.
    • This regularisation technique gives sparse results, which leads to feature selection as well.
  • L2 or Ridge regularisation:
    • Here, the squares of the coefficients are added to the cost function.

How to choose the value of the parameter Learning rate?

  • If the value is too small, the gradient descent algorithm takes ages to converge to the optimal solution.
  • On the other hand, if the value of the learning rate () is high, the gradient descent will overshoot the optimal solution and most likely never converge to the optimal solution.
  • To overcome this problem, you can try different values of alpha over a range of values and plot the cost vs the number of iterations. Then, based on the graphs, the value corresponding to the graph showing the rapid decrease can be chosen.
  • If you see that the cost is increasing with the number of iterations, then your learning rate parameter is high and it needs to be decreased.
More...

Practice questions

06 Jul 2020
statistics practice-questions

What is the difference between univariate, bivariate and multivariate analyses?

  • Univariate analysis: The data which is being analysed contains one variable. For example, analysing the customers visiting an e-commerce fashion website. The data can be divided into men, women and kids, as following:

    Category Number per hour
    Women 78
    Men 49
    Kids 12

    The major purpose of this analysis is to describe and find patterns in data.

  • Bivariate analysis: It is used to find relationships between two different variables. For example, finding the relationship between Employee satisfaction level and income.
  • Multivariate analysis: It is the analysis of three or more variables. For example, analyse the factors that could have affected the lower sales of a newly launched product.
More...