
Amir Abolhassani- PhD
- Senior Data Scientist at General Motors Company
Amir Abolhassani
- PhD
- Senior Data Scientist at General Motors Company
About
13
Publications
29,872
Reads
How we measure 'reads'
A 'read' is counted each time someone views a publication summary (such as the title, abstract, and list of authors), clicks on a figure, or views or downloads the full-text. Learn more
267
Citations
Introduction
Current institution
Additional affiliations
August 2012 - August 2016
Publications
Publications (13)
div>Ridesharing is a shared vehicle service with the potential to meet the growing travel demand and shortage in transportation infrastructure capacity. Ridesharing services reduce the number of vehicles and reduce traffic congestion and emissions while providing mobility services to the same number of people with no additional transportation infra...
This paper addresses the question “Is energy that different from labor?” from the perspective of efficiency. It presents a novel statistical analysis for the auto assembly industry in North America to examine the determinants of relative energy intensity, and contrasts this with a similar analysis of the determinants of another important factor of...
Manufacturing processes usually exhibit mixed operational conditions (OCs) due to changes in process/tool/equipment health status. Undesired OCs are direct causes of out-of-control production and thus need to be identified. Data-driven OC identification has been widely used for recognizing undesired OCs, yet most methods of this kind require labels...
Machine faults and systematic failures are resulted from manufacturing process deteriorations. With early recognition of patterns closely related to process deteriorations, e.g., trends, preventative maintenance can be conducted to avoid severe loss of productivity. Change-point detection identifies the time when abnormal patterns occur, thus are i...
As a critical asset, gantry has wide applications in many fields such as medical image area, infrastructure, and heavy industry. Mostly, the gantry is reliable, however, the loss led by the gantry lockout is inestimable enormous. Moreover, there are limited previous gantry studies concentrate on the statistical quality control to detect the fault n...
Automatic sensing devices and computer systems have been widely adopted by the automotive manufacturing industry, which are capable to record machine status and process parameters nonstop. While a manufacturing process always has natural variations, it is crucial to detect significant changes to the process for quality control, as such changes may...
Even though there is a pressing need for continuous productivity improvement, studies that employ robust empirical analysis of strategies and factors to enhance productivity in the North American Automotive Industry are very scarce. In this study, robust and hybrid models of the most popular productivity measurement in the automotive industry, Hour...
Energy management has become crucial for the industrial sector as a structured approach to lowering the cost of production and in reducing the carbon footprint. With the development of ISO 50001 standard, energy management has enticed the attention of upper level management in terms of continuous improvement. The ISO 50001 standard requires an inte...
The aim of this study is to define a robust estimation model of the most dominant labour productivity measurement, Hours per Vehicle (HPV), in the auto industry. Data utilised in this study were from 10 different multinational North American carmakers from 1999 to 2007. Through a comprehensive literature review and practical consideration, 13 impor...
Purpose
The purpose of this paper is to identify effective factors, their impact, and find estimation models of the most well-known productivity measurement, hours-per-vehicle (HPV), in the automotive industry in North American manufacturing plants.
Design/methodology/approach
Data used in this study were from North American plants that participat...
The research investigates current strategies that help automobile manufacturers to enhance their productivity. The study utilizes robust statistical methods to define the most important factors on the Hours per Vehicle (HPV) in the automotive industry in North American manufacturing plants. Data are synthesized using a uniform methodology from info...
Purpose
The purpose of this paper is to analyze lean strategic practices being implemented in manufacturing facilities throughout Pennsylvania and West Virginia and identify the difficulty in implementing those lean practices.
Design/methodology/approach
A skip logic questionnaire was developed into multiple sections for analysis; demographics of...
Questions
Questions (19)
What is the right way of interpreting the relation between Y & X in the following graphs? I have calculated the Pearson and Spearman correlation. However, they are significant, but the correlation coefficient is too small (There is a high number of observations around zero on the X axis, I'm wondering if I need to use the dataset as it is or do something specific prior calculating the correlation coefficient between Y & X). Thank you.
I have a conceptual question related to fuzzy logic method. I have a fuzzy environment and planning to use a fuzzy logic method to define/develop a productivity measure. However, I’m confused about the relation between the sample size (with predefined range for each input/feature which will limit the inputs range) in the data and fuzzy logic method parameters such as membership function etc. To me it seems there is no relationship, if you have one observation or 1000, you can define the Fuzzy classification of input and output variables/levels (with predefined range which the range is not coming from the data) and the functions, no matter how many observations are available. In that case, the surface viewer is something that is identified by the rules only and not available data (no matter if there is one observation or 1000 ones) and the hypothetical values of a use case that you are identifying.
Am I right or I'm missing something? Thank you.
P.S. I have attached a surface viewer example.
Hello, There is a dataset with several KPIs which are varying between (0,1). What is the best analytical approach to split the data and define a line in two dimensional (or define a plane in multi-dimensional space) based on data behavior and practical assumptions/considerations (there is some recommended ranges for each KPIs etc.)?
For instance in the attached screenshot, I want to flag the individuals/observations in Ae area for more investigation. I want to be able to apply the proposed approach in multi-dimensional space with several KPIs as well. Any thoughts would be appreciated.
Imagine that I have a regression model equation as,
Y= B0 + B1X1 + B2X2 + B3X3 + B4X4 + e
Where,
X1 is a variable with continuous value
X2 is a categorical variable at two levels such as gender
X3 is a variable with integer value
X4 is another categorical variable at two levels such as if a person is experienced or inexperience
I can solve it by using an OLS regression or using an unbalanced panel data. In the OLS regression, I can compare male group verses female group or experience group versus inexperience group which seems like I'm treating them as fixed effects in an unbalanced panel data. In that case, both of them are the same and I have to get the same results, am I right? Is there any advantages using OLS regression versus unbalanced panel data? Can I have a panel data without considering the time impact on the dependent variable? Thank you.
In one hand, there are limited number of data-points and so many features for the analysis, let’s say 50 data-points and 60 features. On the other hand, the unusual data-points are more important than the usual-points and you want to know the underline facts of happening those unusual data-points. How do you approach the analysis to tackle these two issues at the same time?
Any help would be appreciated.
I'm interested in knowing the limitations of different open-source solvers such as PuLP, GLPK, and Symphony comparing to commercial optimization solvers such as CPLEX or Gurobi.
Any information would be appreciated.
I would appreciate if you could define what different steps are necessary to develop a policy for determining the frequency of a PMs on a specific equipment. For instance, we have a pump that it has three different PMs and we are not gathering any sensors' information such as current, vibration, etc. How can we define the frequency of these PMs for this pump that might have various failure modes? Thank you.
There are 11 predictors, combination of continuous and categorical variables. The response is a continuous variable and there isn’t any multicollinearity between variables, all VIFs are lower than 2.9. Multiple linear regression, ridge, lasso, and weighted regression methods (the weights are calculated through a robust method) are used. In contrary to other regression methods, the sign of a particular predictor coefficients, a binary variable, is flipping when weighted regression method is used.
I appreciate for any feedback and thought about the flip sign of the binary variable.