Skip to Content

Krenicki Center

Past Projects

Our Capabilities: Interdisciplinary Teams Utilize Diverse Skillsets

Business Skills

Finance
Marketing
Retail
Supply Chain
Manufacturing
Engineering

Technical Skills

Data Science, AI/ML, Predicitive Modeling
Simulation, Optimization
Image, Audio, and Natural Language Processing
Web Crawling, Automation
Unstructured Data Analysis
Data Pipelining and Architecting
Analytical, Visualization, and Cloud Platforms

Behavioral Skills

Project Management
Process and Solution Design
Dashboarding
Presentation Delivery
Leadership and Staff Education

Past Projects

Interpretation of Forecasting Models

  1. Objectives and Methodology
    1. To explore different algorithms that can be used for interpretation of the workings behind black box models
    2. To study the contribution and impact of each input variable on the forecast prediction provided by the black box model
    3. To create approximation models to get above outputs for single observation and observations over a period
  2. Data and overview of experiments
    1. Out of ~14k unique DFUs and 24 input features, top 10 DFUs and 4 input features wereused for the final model building process
    2. Based on business requirements of the output, XGBoost interpretation using Linear Regression and LIME; and Prophet were selected as final methods for dashboard
    3. Global surrogate models explain holistic variable contributions and impacts from the black box model for the entire dataset
    4. Whereas local surrogate models explain the variable contribution for a single test observation
  3. XG Boost – Global Explanations
    1. Forecasts from the XGBoost model are used as target variable for the linear regression that gives a global interpretation of how XGBoost works under the hood
    2. First level of interpretation can be done using significance to understand the drivers of the forecast; sign and magnitude of the coefficient can then serve to give a direction for marketing strategy
  4. XG Boost – Local Explanations
    1. Total Sales(LIME) = Intercept + Total Points of Distribution Total + Depth of Discount + Month + Day
    2. Contribution by seasonality to product sales increases on Sept 21
  5. Prophet – Local and Global Explanations
    1. Prophet Model: F(t) + F(s) + holidays + extra regressors ( Total Points of Distribution + Depth of Discount)
    2. Sales Forecast = Day of Month + Discount + Holiday + Month + TPD + Year
  6. Conclusion and Future Scope
    1. This study tries to create an approximate model around the original forecasting model in order to give users a sense of how different parameters such as discount, trend, seasonality are affecting the sales of their product at different points in time.
    2. Checking the statistical significance of the variable and visualizing the contributions can help understand the business impact from the variable
    3. Combining Prophet + LIME Models explains and supports feature contribution on sales and business critical values
    4. Not all variables for all products will have similar impacts on sales and for few products there aren't much explanations with the given variables.
    5. Since interpretation models are approximations of black box models, the exact contribution cannot be directly inferred. However, it is useful to indicate the direction and magnitude of the impact.

Forecasting and Methodology: Improving the Accuracy of Predicted Ocean Travel Time

An American exercise equipment and media company had received a record-breaking order volume alongside a reduction in staff due to the COVID-19.​

The supply chain was broken, and delivery times changed from 30 to 60 days due to anchoring and port congestion​. Our objective was to:

  • Identify the factors affecting ocean travel time i.e., from point A to B.​
  • Improve the current forecasting methodology using external factors like marine traffic and port congestion.​

Methodology

The client presented their forecasting model, wherein ocean lead time was factored using simple exponential smoothing with no trend, seasonality and external factors​.

Validation WAPE (Weighted Absolute Percentage Error) of 11.35% for data from week of 1st February 2021 to week of 17th May 2021 had been achieved till our team was brought in.

Our approach to problem solving involved the following steps:

Data Collection: Collect the Anchorage Time, Port Time, No. of Vessels, and No. of Calls at the port of arrival for each week.

Data Exploration: Remove “Calls” cluster of variables using correlation analysis.

Feature Engineering: Create lagged derivatives for Ocean, Anchor, Port & Vessels clusters.

Data Partition: Split the Train & Test data using rolling forecasting origin techniques.

Forecasting: Build regression, SVM, Random Forest, Gradient Boosting models.

Outcome

We achieved a validation WAPE of 11.13% and were able to demonstrate that among the external factor clusters - Anchor, Port, Vessels and Lead Times; Anchor variables are not significant in the presence of the other three clusters.​

Apart from the Historical Lead Times in the Ocean, the previous week’s Port Time & Number of vessels 2-weeks earlier are affecting the Lead Time Ocean for the current week. ​

Tools used

Python & Tableau

Redesigning a Value Based Program

In the collaboration project with Accenture and Krenicki Center, students assisted in developing semantic layer for reports, developed report content, prepared test data, and assisted in building data extracts. The final deliverables were 5 reporting dashboards successfully integrated with the client’s database. The students undergo three main phases of the project to achieve the final deliverables:

  1. Developing wireframes with test data
  2. Connecting reports to database
  3. Querying data
Throughout the project, students learned technical skills such as data visualization with Tableau Desktop and SQL querying with QMF. Moreover, students also strengthened soft skills by building strong communication and presentation skills through weekly meetings with the Accenture team.

Reducing the Uncertainty During a SAP Transition

Goals:

  1. Create an interactive dashboard to help team view insights and tag points of interest.
  2. Leverage predictive analytics and forecasting to decrease uncertainty in SAP transition.
  3. Use our creativity and analytical mindset to see where other improvements could be made based on the suggestions of Accenture leadership. a. This was continuously evolving, but the overall goal – since no specific SOW was used for the majority of the project.

Processes:

  1. Dashboard a. Learning Power BI and utilizing Data Visualization best practices to create a final product for the client.
    1. Pre-processing large and unruly data sets to make understandings allowed by the visualization easier to attain.
    2. Solidifying front and backend processes in order to make sure the dashboard would function across different aspects of the business.
  2. Forecasting a. Immense pre-processing to be able to feed the data to well known forecasting engines in R (i.e. ETS).
    1. Running the engines and working to increase accuracy with limited data and loose requirements for accuracy.
    2. With the forecasted data, making the output file compatible with visualization tools and techniques.
  3. Creation past the original SOW a. Present to the Accenture team the work we had done to create a jumping point for a brainstorming session.
    1. Brainstorming session with Accenture team in Chicago office to direct the remainder of our time with the project.
    2. Splitting up the brainstormed ideas into four subsequent projects i. Staffing Plan analysis
    3. Meeting Cadence Trends
    4. Peaks & Valley Analysis
    5. Data Hygiene
    6. Devising the analytical methods to attack the problems outlined above.
    7. Working through those sub-projects and presenting at least once a month for Accenture team feedback
    8. Collating all the above work for a final presentation with Accenture Leadership and the team that is taking over the work from here on out.

Commonly Requested Projects

Exploring ways to streamline operations and reduce manual work through automation and workflow analysis.

  • Automating intake or application processes
  • Identifying bottlenecks in internal workflows
  • Digitizing paper-based or spreadsheet-driven systems

Applying machine learning to forecast, classify, or support smarter decisions.

  • Forecasting sales or resource needs
  • Predicting customer churn or donor behavior
  • Prototyping recommendation engines or classification tools

Designing dashboards to track KPIs, monitor trends, and support decision-making.

  • Creating real-time dashboards in Power BI or Tableau
  • Visualizing customer engagement, financials, or performance metrics
  • Building tools for internal or external reporting needs

Using data to uncover patterns, inform strategy, and support customer engagement.

  • Segmenting customers based on behavior or demographics
  • Analyzing survey data to identify key themes or trends
  • Assessing campaign effectiveness or market potential

Improving how data is organized, accessed, and used for regular reporting.

  • Streamlining data entry and reporting workflows
  • Enhancing data readiness for future analytics work
  • Standardizing reporting tools across teams or platforms

Research Abstracts

(This list is not comprehensive)

Project Description

This project, supported by the Krenicki Center for Business Analytics and Machine Learning, aims to develop an open-source logistics simulator using SimPy. The simulator will serve as a comprehensive tool for researchers, educators, and industry professionals to model and analyze complex logistics systems. Two students, one from the Department Computer Science and one from the Daniels School of Business, will collaborate to design, develop, and test the simulator under the mentorship Prof. Mohit Tawarmalani and Prof. Bruno Ribiero.

Project Objectives

  • Design and implement an open-source logistics simulator leveraging SimPy.
  • Develop a user-friendly interface and documentation for community adoption.
  • Engage with the open-source community to facilitate broader adoption and contribution.

This study examines audit firms’ use of social media around reputation-damaging events. Using data extracted from the Twitter accounts of Big 4 audit firms, we find audit firms increase their social media activity in the three days surrounding the announcement of a client restatement or the loss of a client. In cross-sectional analyses, we find that heightened social media activity is significantly more pronounced for restatements that have an impact on net income and the loss of major clients. Additionally, we find that strategic social media activity is concentrated among audit firms with relatively higher PCAOB inspection deficiency rates, suggesting that heightened social media activity surrounding reputation-damaging events is driven by reputation concerns. Finally, we also find evidence that increased social media activity is positively associated with audit firm reputation among Twitter users and audit clients.

Street parking offers convenience for customers driving to their favorite retail destinations. Yet, city officials often raise parking fees in an effort to reduce traffic congestion and enhance urban living. The impact of these fee hikes on local retailers remains ambiguous. On one hand, higher parking costs might deter potential customers from visiting the retail area; on the other, improved traffic flow and more efficient parking turnover could enhance consumer access. Using geolocation data and leveraging a recent policy introduced by the City of Chicago, we measure the impact of an increase in parking fees on consumer footfall at local restaurants. We find that a small increase in hourly street parking fees reduces the average footfall at restaurants without customer parking by as much as 15%, while boosting average footfall at nearby competing restaurants that offer parking by around 7%. In contrast, we do not find evidence of any positive effect of the parking fee hike on overall restaurant footfall. Our estimates also indicate that this policy change disproportionately affects the economically poorer sections of the population, reducing their access to these restaurants by twice as much compared to other groups.