You are excited. You have got that much awaited interview call for that dream analytics job. You are confident you will be perfect for the job. Now all that remains is convincing the interviewer. Don’t you wish you knew what kind of questions they are going to be ask?
As co founder and one of the chief trainers at Jigsaw Academy, an online analytics training institute, I regularly get calls from our students days before their scheduled interview asking me just this. I am going to share with you just what I share with them. Here you go. Below are a few of the more popular questions you could get asked and the corresponding answers in a nutshell.
Broadly speaking these are the steps. Of course these may vary slightly depending on the type of problem, data, tools available etc.
1. Problem definition – The first step is to of course understand the business problem. What is the problem you are trying to solve – what is the business context? Very often however your client may also just give you a whole lot of data and ask you to do something with it. In such a case you would need to take a more exploratory look at the data. Nevertheless if the client has a specific problem that needs to be tackled, then then first step is to clearly define and understand the problem. You will then need to convert the business problem into an analytics problem. I other words you need to understand exactly what you are going to predict with the model you build. There is no point in building a fabulous model, only to realise later that what it is predicting is not exactly what the business needs.
2. Data Exploration – Once you have the problem defined, the next step is to explore the data and become more familiar with it. This is especially important when dealing with a completely new data set.
Over 100,000 people subscribe to our newsletter.
See stories of Analytics and AI in your inbox.
3. Data Preparation – Now that you have a good understanding of the data, you will need to prepare it for modelling. You will identify and treat missing values, detect outliers, transform variables, create binary variables if required and so on. This stage is very influenced by the modelling technique you will use at the next stage. For example, regression involves a fair amount of data preparation, but decision trees may need less prep whereas clustering requires a whole different kind of prep as compared to other techniques.
4. Modelling – Once the data is prepared, you can begin modelling. This is usually an iterative process where you run a model, evaluate the results, tweak your approach, run another model, evaluate the results, re-tweak and so on….. You go on doing this until you come up with a model you are satisfied with or what you feel is the best possible result with the given data.
5. Validation – The final model (or maybe the best 2-3 models) should then be put through the validation process. In this process, you test the model using completely new data set i.e. data that was not used to build the model. This process ensures that your model is a good model in general and not just a very good model for the specific data earlier used (Technically, this is called avoiding over fitting)
6. Implementation and tracking – The final model is chosen after the validation. Then you start implementing the model and tracking the results. You need to track results to see the performance of the model over time. In general, the accuracy of a model goes down over time. How much time will really depend on the variables – how dynamic or static they are, and the general environment – how static or dynamic that is.
Data exploration is done to become familiar with the data. This step is especially important when dealing with new data. There are a number of things you will want to do in this step –
a. What is there in the data – look at the list of all the variables in the data set. Understand the meaning of each variable using the data dictionary. Go back to the business for more information in case of any confusion.
b. How much data is there – look at the volume of the data (how many records), look at the time frame of the data (last 3 months, last 6 months etc.)
c. Quality of the data – how much missing information, quality of data in each variable. Are all fields usable? If a field has data for only 10% of the observations, then maybe that field is not usable etc.
d. You will also identify some important variables and may do a deeper investigation of these. Like looking at averages, min and max values, maybe 10th and 90th percentile as well…
e. You may also identify fields that you need to transform in the data prep stage.
In data preparation, you will prepare the data for the next stage i.e. the modelling stage. What you do here is influenced by the choice of technique you use in the next stage.
But some things are done in most cases – example identifying missing values and treating them, identifying outlier values (unusual values) and treating them, transforming variables, creating binary variables if required etc,
This is the stage where you will partition the data as well. i.e create training data (to do modelling) and validation (to do validation).
The first step is to identify variables with missing values. Assess the extent of missing values. Is there a pattern in missing values? If yes, try and identify the pattern. It may lead to interesting insights.
If no pattern, then we can either ignore missing values (SAS will not use any observation with missing data) or impute the missing values.
Simple imputation – substitute with mean or median values
Case wise imputation –for example, if we have missing values in the income field.
You can identify outliers using graphical analysis and univariate analysis. If there are only a few outliers, you can assess them individually. If there are many, you may want to substitute the outlier values with the 1stpercentile or the 99th percentile values.
If there is a lot of data, you may decide to ignore records with outliers.
Not all extreme values are outliers. Not all outliers are extreme values.
You can use different methods to assess how good a logistic model is.
a. Concordance – This tells you about the ability of the model to discriminate between the event happening and not happening.
b. Lift – It helps you assess how much better the model is compared to random selection.
c. Classification matrix – helps you look at the false positives and true negatives.
Some other general questions you will most likely be asked:
- What have you done to improve your data analytics knowledge in the past year?
- What are your career goals?
- Why do you want a career in data analytics?
The answers to these questions will have to be unique to the person answering it. The key is to show confidence and give well thought out answers that demonstrate you are knowledgeable about the industry and have the conviction to work hard and excel as a data analyst.
Discover relevant questions―and detailed answers―to help you prepare for job interviews and break into the field of analytics. This book contains more than 300 questions based on consultations with hiring managers and technical professionals already working in analytics. Interview Questions in Business Analytics fills a gap in information on business analytics for job seekers.
Get your copy at: http://www.amazon.in/Interview-Questions-Business-Analytics-Bhasker/dp/1484206002