UC Business Analytics R Programming Guide2020-03-04T22:25:39+00:00http://uc-r.github.ioCreating text features with bag-of-words, n-grams, parts-of-speach and more2018-10-02T00:00:00+00:00http://uc-r.github.io/2018/10/02/creating-text-features-with-bag-of-words-and-more<p><img src="/public/images/analytics/feature-engineering/text-features.jpg" style="float:right; margin: 0px -5px 0px 10px; width: 35%; height: 35%;" />
Historically, data has been available to us in the form of numeric (i.e. customer age, income, household size) and categorical features (i.e. region, department, gender). However, as organizations look for ways to collect new forms of information such as unstructured text, images, social media posts, etcetera, we need to understand how to convert this information into structured features to use in data science tasks such as customer segmentation or prediction tasks. In this <a href="http://uc-r.github.io/creating-text-features">post</a>, we explore a few fundamental feature engineering approaches that we can start using to convert unstructured text into structured features.</p>
Multivariate Adaptive Regression Splines2018-09-08T00:00:00+00:00http://uc-r.github.io/2018/09/08/mars<p><img src="/public/images/analytics/mars/3D_Mars.png" style="float:right; margin: 2px 0px 0px 10px; width: 25%; height: 25%;" />
Several previous tutorials (i.e. linear regression, logistic regression, regularized regression) discussed algorithms that are intrinsically linear. Many of these models can be adapted to nonlinear patterns in the data by manually adding model terms (i.e. squared terms, interaction effects); however, to do so you must know the specific nature of the nonlinearity a priori. Alternatively, there are numerous algorithms that are inherently nonlinear. When using these models, the exact form of the nonlinearity does not need to be known explicitly or specified prior to model training. Rather, these algorithms will search for, and discover, nonlinearities in the data that help maximize predictive accuracy. This <a href="http://uc-r.github.io/mars">latest tutorial</a> discusses multivariate adaptive regression splines (MARS), an algorithm that essentially creates a piecewise linear model which provides an intuitive stepping block into nonlinearity after grasping the concept of linear regression and other intrinsically linear models.</p>
Interpreting Machine Learning Models with the iml Package2018-08-01T00:00:00+00:00http://uc-r.github.io/2018/08/01/iml-pkg<p><img src="/public/images/analytics/ML_interpretation/iml_icon.jpg" style="float:right; margin: 0px 0px 0px 0px; width: 30%; height: 30%;" /></p>
<p>With machine learning interpretability growing in importance, several R packages designed to provide this capability are gaining in popularity. In recent blog posts I assessed <a href="http://uc-r.github.io/lime"><code class="language-plaintext highlighter-rouge">lime</code></a> for model agnostic local interpretability functionality and <a href="http://uc-r.github.io/dalex"><code class="language-plaintext highlighter-rouge">DALEX</code></a> for both local and global machine learning explanation plots. This <a href="http://uc-r.github.io/iml-pkg">newest tutorial</a> examines the <code class="language-plaintext highlighter-rouge">iml</code> package to assess its functionality in providing machine learning interpretability to help you determine if it should become part of your preferred machine learning toolbox.</p>
Model Interpretability with DALEX2018-07-11T00:00:00+00:00http://uc-r.github.io/2018/07/11/dalex<p><img src="/public/images/analytics/ML_interpretation/dalex_logo.png" style="float:right; margin: 0px 0px 0px 10px; width: 35%; height: 35%;" /></p>
<p>As advanced machine learning algorithms are gaining acceptance across many organizations and domains, machine learning interpretability is growing in importance to help extract insight and clarity regarding how these algorithms are performing and why one prediction is made over another. There are many methodologies to interpret machine learning results (i.e. variable importance via permutation, partial dependence plots, local interpretable model-agnostic explanations), and many machine learning R packages implement their own versions of one or more methodologies. However, some recent R packages that focus purely on ML interpretability agnostic to any specific ML algorithm are gaining popularity. One such package is <code class="language-plaintext highlighter-rouge">DALEX</code> and this <a href="http://uc-r.github.io/dalex">latest tutorial</a> covers what this package does (and does not do) so that you can determine if it should become part of your preferred machine learning toolbox.</p>
Gradient Boosting Machines2018-06-14T00:00:00+00:00http://uc-r.github.io/2018/06/14/gbm-regression<p><img src="/public/images/analytics/gbm/boosted_stumps.gif" style="float:right; margin: 2px 5px 0px 20px; width: 30%; height: 30%;" />
Gradient boosting machines (GBMs) are an extremely popular machine learning algorithm that have proven successful across many domains and is one of the leading methods for winning Kaggle competitions. Whereas <a href="random_forests">random forests</a> build an ensemble of deep independent trees, GBMs build an ensemble of shallow and weak successive trees with each tree learning and improving on the previous. When combined, these many weak successive trees produce a powerful “committee” that are often hard to beat with other algorithms. This <a href="http://uc-r.github.io/gbm_regression">latest tutorial</a> covers the fundamentals of GBMs for regression problems.</p>
Visualizing ML Models with LIME2018-06-08T00:00:00+00:00http://uc-r.github.io/2018/06/08/lime<p><img src="https://www.data-imaginist.com/assets/images/lime_logo_small.jpg" style="float:right; margin: 0px 0px 0px 0px; width: 45%; height: 45%;" />
Machine learning (ML) models are often considered “black boxes” due to their complex inner-workings. More advanced ML models such as random forests, gradient boosting machines (GBM), artificial neural networks (ANN), among others are typically more accurate for predicting nonlinear, faint, or rare phenomena. Unfortunately, more accuracy often comes at the expense of interpretability, and interpretability is crucial for business adoption, model documentation, regulatory oversight, and human acceptance and trust. Luckily, several advancements have been made to aid in interpreting ML models. This <a href="http://uc-r.github.io/lime">latest tutorial</a> demonstrates how to use the <code class="language-plaintext highlighter-rouge">lime</code> package, which helps to perform local interpretations of ML models.</p>
Random Forests2018-05-09T00:00:00+00:00http://uc-r.github.io/2018/05/09/random-forests<p><img src="/public/images/analytics/random_forests/RF_icon.jpg" style="float:right; margin: 2px 5px 0px 20px; width: 30%; height: 30%;" /></p>
<p>Bagging regression trees is a technique that can turn a single tree model with high variance and poor predictive power into a fairly accurate prediction function. Unfortunately, bagging regression trees typically suffers from tree correlation, which reduces the overall performance of the model. <strong><em>Random forests</em></strong> are a modification of bagging that builds a large collection of <em>de-correlated</em> trees and have become a very popular “out-of-the-box” learning algorithm that enjoys good predictive performance. This <a href="http://uc-r.github.io/random_forests">latest tutorial</a> will cover the fundamentals of random forests.</p>
Regression Trees2018-04-28T00:00:00+00:00http://uc-r.github.io/2018/04/28/regression-trees<p><img src="/public/images/analytics/regression_trees/iris.png" style="float:right; margin: 0px 0px 0px 0px; width: 30%; height: 30%;" />
Basic <strong><em>regression trees</em></strong> partition a data set into smaller groups and then fit a simple model (constant) for each subgroup. Unfortunately, a single tree model tends to be highly unstable and a poor predictor. However, by bootstrap aggregating (<strong><em>bagging</em></strong>) regression trees, this technique can become quite powerful and effective. Moreover, this provides the fundamental basis of more complex tree-based models such as <em>random forests</em> and <em>gradient boosting machines</em>. This <a href="http://uc-r.github.io/regression_trees">latest tutorial</a> will get you started with regression trees and bagging.</p>
Naïve Bayes Classifier2018-04-20T00:00:00+00:00http://uc-r.github.io/2018/04/20/naive-bayes<p><img src="/public/images/analytics/naive_bayes/naive_bayes_icon.png" style="float:right; margin: 10px 0px 0px 0px; width: 50%; height: 50%;" /></p>
<p>The <strong><em>Naïve Bayes classifier</em></strong> is a simple probabilistic classifier which is based on Bayes theorem but with strong assumptions regarding independence. Historically, this technique became popular with applications in email filtering, spam detection, and document categorization. Although it is often outperformed by other techniques, and despite the naïve design and oversimplified assumptions, this classifier can perform well in many complex real-world problems. And since it is a resource efficient algorithm that is fast and scales well, it is definitely a machine learning algorithm to have in your toolkit. This <a href="http://uc-r.github.io/naive_bayes">tutorial</a> will introduce you to this simple classifier.</p>
Feedforward Deep Learning Models2018-04-09T00:00:00+00:00http://uc-r.github.io/2018/04/09/feedforward-deep-models<p><img src="/public/images/analytics/deep_learning/deep_nn.png" style="float:right; margin: 2px 0px 0px 10px; width: 55%; height: 55%;" /></p>
<p>Machine learning algorithms typically search for the optimal representation of data using some feedback signal (aka objective/loss function). However, most machine learning algorithms only have the ability to use one or two layers of data transformation to learn the output representation. As data sets continue to grow in the dimensions of the feature space, finding the optimal output representation with a <em>shallow</em> model is not always possible. Deep learning provides a multi-layer approach to learn data representations, typically performed with a <em>multi-layer neural network</em>. Like other machine learning algorithms, deep neural networks (DNN) perform learning by mapping features to targets through a process of simple data transformations and feedback signals; however, DNNs place an emphasis on learning successive layers of meaningful representations. Although an intimidating subject, the overarching concept is rather simple and has proven highly successful in predicting a wide range of problems (i.e. image classification, speech recognition, autonomous driving). This <a href="http://uc-r.github.io/feedforward_DNN">tutorial</a> will teach you the fundamentals of building a <em>feedfoward</em> deep learning model.</p>
Regularized Regression2018-03-28T00:00:00+00:00http://uc-r.github.io/2018/03/28/regularized-regression<p><img src="/public/images/analytics/regularized_regression/regularization_logo.png" style="float:right; margin: 0px 0px 0px 0px; width: 50%; height: 50%;" />
Linear regression is a simple and fundamental approach for supervised learning. Moreover, when the assumptions required by ordinary least squares (OLS) regression are met, the coefficients produced by OLS are unbiased and, of all unbiased linear techniques, have the lowest variance. However, in today’s world, data sets being analyzed typically have a large amount of features. As the number of features grow, our OLS assumptions typically break down and our models often overfit (aka have high variance) to the training sample, causing our out of sample error to increase. <strong><em>Regularization</em></strong> methods provide a means to control our regression coefficients, which can reduce the variance and decrease out of sample error. This latest <a href="http://uc-r.github.io/regularized_regression">tutorial</a> explains why you should know this technique and how to implement it in R.</p>
Visual Data Exploration2018-01-22T00:00:00+00:00http://uc-r.github.io/2018/01/22/visual-data-exploration<p><img src="/public/images/visual/graphical_data_analysis/gda_icon.png" style="float:right; margin: 20px 0px 0px 5px; width: 50%; height: 50%;" /></p>
<p>Visual data exploration is a mandatory intial step whether or not more formal analysis follows. When combined with <a href="http://uc-r.github.io/descriptive">descriptive statistics</a>, visualization provides an effective way to identify summaries, structure, relationships, differences, and abnormalities in the data. Often times no elaborate analysis is necessary as all the important conclusions required for a decision are evident from simple visual examination of the data. Other times, data exploration will be used to help guide the data cleaning, feature selection, and sampling process. Regardless, visual data exploration is about investigating the characteristics of your data set. To do this, we typically create numerous plots in an interactive fashion. This <a href="http://uc-r.github.io/gda">tutorial</a> will show you how to create plots that answer some of the fundamental questions we typically have of our data.</p>
Exponential Smoothing Models2017-10-16T00:00:00+00:00http://uc-r.github.io/2017/10/16/exponential-smoothing<p><img src="/public/images/analytics/time_series/es10-1.png" style="float:right; margin: 2px 0px 0px 10px; width: 40%; height: 40%;" />
Exponential forecasting models are smoothing methods that have been around since the 1950s and are extremely effective. Where <a href="http://uc-r.github.io/ts_benchmarking#naive">niave forecasting</a> places 100% weight on the most recent observation and <a href="http://uc-r.github.io/ts_moving_averages">moving averages</a> place equal weight on <em>k</em> values, exponential smoothing allows for weighted averages where greater weight can be placed on recent observations and lesser weight on older observations. Exponential smoothing methods are intuitive, computationally efficient, and generally applicable to a wide range of time series. Consequently, exponentially smoothers are great forecasting tools to have and <a href="http://uc-r.github.io/ts_exp_smoothing">this tutorial</a> will walk you through the basics.</p>
Support Vector Machines2017-09-27T00:00:00+00:00http://uc-r.github.io/2017/09/27/support-vector-machines<p><img src="/public/images/analytics/svm/unnamed-chunk-13-1.png" style="float:right; margin: 0px -5px 0px 10px; width: 45%; height: 45%;" />
The advent of computers brought on rapid advances in the field of statistical classification, one of which is the <em>Support Vector Machine</em>, or SVM. The goal of an SVM is to take groups of observations and construct boundaries to predict which group future observations belong to based on their measurements. The different groups that must be separated will be called “classes”. SVMs can handle any number of classes, as well as observations of any dimension. SVMs can take almost any shape (including linear, radial, and polynomial, among others), and are generally flexible enough to be used in almost any classification endeavor that the user chooses to undertake. This new <a href="http://uc-r.github.io/svm">tutorial</a> will introduce you to this supervised classification technique.</p>
Introduction to Neural Networks2017-09-07T00:00:00+00:00http://uc-r.github.io/2017/09/07/Introduction-to-Neural-Networks<p><img src="https://upload.wikimedia.org/wikipedia/commons/6/60/ArtificialNeuronModel_english.png" style="float:right; margin: 0px -5px 0px 10px; width: 45%; height: 45%;" />
Artificial neural networks (ANNs) describe a specific class of machine learning algorithms designed to acquire their own knowledge by extracting useful patterns from data. ANNs are function approximators, mapping inputs to outputs, and are composed of many interconnected computational units, called neurons. Each individual neuron possesses little intrinsic approximation capability; however, when many neurons function cohesively together, their combined effects show remarkable learning performance. A new set of <a href="http://uc-r.github.io/predictive#deep-learning">tutorials</a> introduce you to this popular analytic technique.</p>
Imprecise Regression2017-08-23T00:00:00+00:00http://uc-r.github.io/2017/08/23/imprecise-regression<p><img src="/public/images/analytics/imprecise_regression/unnamed-chunk-3-1.png" style="float:right; margin: 0px -5px 0px 10px; width: 45%; height: 45%;" />
Imprecise regression is a generalization of linear regression that gives us stronger tools for modeling uncertainty. In a typical linear regression setting, we consider the input data to be precise observations: single points. In imprecise regression, we generalize the notion of observations to intervals rather than points. This allows us to more accurately represent scenarios with measurement error or other sources of uncertainty that make our input data “fuzzy”. This <a href="http://uc-r.github.io/imprecise_regression">tutorial</a> is based on the imprecise regression work in <a href="http://www.sciencedirect.com/science/article/pii/S0888613X12000862">Cattaneo and Wiencierz (2012)</a> and will teach you the fundamentals of this technique.</p>
Moving Averages2017-08-04T00:00:00+00:00http://uc-r.github.io/2017/08/04/moving-averages<p><img src="/public/images/analytics/time_series/ma_icon.png" style="float:right; margin: 2px 0px 0px 10px; width: 40%; height: 40%;" />
Smoothing methods are a family of forecasting approaches that average values over multiple periods in order to reduce the noise and uncover patterns in the data. Moving averages are one such smoothing method. In this new <a href="http://uc-r.github.io/ts_moving_averages">time series tutorial</a>, you will learn the basics of performing smoothing averages.</p>
Benchmark Methods & Forecast Accuracy2017-06-16T00:00:00+00:00http://uc-r.github.io/2017/06/16/benchmarking-methods-and-forecast-accuracy<p><img src="/public/images/analytics/time_series/partitioning-1.png" style="float:right; margin: 2px 0px 0px 10px; width: 40%; height: 40%;" />
In this new <a href="http://uc-r.github.io/ts_benchmarking">time series tutorial</a>, you will learn general tools that are useful for many different forecasting situations. It will describe some methods for benchmark forecasting, methods for checking whether a forecasting model has adequately utilized the available information, and methods for measuring forecast accuracy. These are important tools to have in your forecasting toolbox as each are leveraged repeatedly as you develop and explore a range of forecasting methods.</p>
New Tutorial on Exploring and Visualizing Time Series2017-06-02T00:00:00+00:00http://uc-r.github.io/2017/06/02/exploring-and-visualizing-time-series<p><img src="/public/images/analytics/time_series/unnamed-chunk-6-1.png" style="float:right; margin: 2px 0px 0px 10px; width: 40%; height: 40%;" />
Time series forecasting is performed in nearly every organization that works with quantifiable data. Retail stores forecast sales. Energy companies forecast reserves, production, demand, and prices. Educational institutions forecast enrollment. Goverments forecast tax receipts and spending. International financial organizations forecast inflation and economic activity. The list is long but the point is short - forecasting is a fundamental analytic process in every organization. This <a href="http://uc-r.github.io/ts_exploration">new tutorial</a> gets you started doing some fundamental time series exploration and visualization.</p>
New Tutorial on Linear Model Selection2017-04-21T00:00:00+00:00http://uc-r.github.io/2017/04/21/model-selection<p><img src="/public/images/analytics/model_selection/unnamed-chunk-6-1.png" style="float:right; margin: 2px 0px 0px 10px; width: 50%; height: 50%;" /></p>
<p>It is often the case that some or many of the variables used in a multiple regression model are in fact <em>not</em> associated with the response variable. Including such irrelevant variables leads to unnecessary complexity in the resulting model. Unfortunately, manually filtering through and comparing regression models can be tedious. Luckily, several approaches exist for automatically performing feature selection or variable selection — that is, for identifying those variables that result in superior regression results. This latest <a href="http://uc-r.github.io/model_selection">tutorial</a> covers a traditional approach known as <em>linear model selection</em>.</p>
New Tutorial Providing an Introduction to `ggplot2`2017-04-07T00:00:00+00:00http://uc-r.github.io/2017/04/07/ggplot-intro<p><img src="http://bradleyboehmke.github.io/public/images/tufte/unnamed-chunk-1-1.png" style="float:right; margin: 2px 0px 0px 2px; width: 50%; height: 50%;" /></p>
<p>Although R provides built-in plotting functions, <code class="language-plaintext highlighter-rouge">ggplot2</code> has become the preeminent visualization package in R. <code class="language-plaintext highlighter-rouge">ggplot2</code> implements the Grammar of Graphics theory making it particularly effective for constructing visual representations of data and learning this library will allow you to make nearly any kind of (static) data visualization, customized to your exact specifications. This <a href="http://uc-r.github.io/ggplot_intro">intro tutorial</a> will get you started in making effective visualizations with R.</p>
New Tutorial on Resampling Methods2017-03-17T00:00:00+00:00http://uc-r.github.io/2017/03/17/resampling<p><img src="/public/images/analytics/resampling/bootstrap.png" style="float:right; margin: 2px 0px 0px 10px; width: 40%; height: 40%;" /></p>
<p>See the new tutorial on <a href="http://uc-r.github.io/resampling_methods">resampling methods</a>, which are an indispensable tool in modern statistics. They involve repeatedly drawing samples from a training set and refitting a model of interest on each sample in order to obtain additional information about the fitted model. For example, in order to estimate the variability of a linear regression fit, we can repeatedly draw different samples from the training data, fit a linear regression to each new sample, and then examine the extent to which the resulting fits differ. Such an approach may allow us to obtain information that would not be available from fitting the model only once using the original training sample.</p>
New Tutorials on Clustering2017-03-03T00:00:00+00:00http://uc-r.github.io/2017/03/03/clustering<p><img src="/public/images/analytics/clustering/kmeans/unnamed-chunk-18-1.png" style="float:right; margin: 2px 0px 0px 10px; width: 40%; height: 40%;" /></p>
<p>Clustering is a broad set of techniques for finding subgroups of observations within a data set. When we cluster observations, we want observations in the same group to be similar and observations in different groups to be dissimilar. Clustering allows us to identify which observations (i.e. customers, students, states) are alike, and potentially categorize them therein. Check out our new tutorials covering <a href="http://uc-r.github.io/kmeans_clustering">k-means</a> and <a href="http://uc-r.github.io/hc_clustering">hierarchical</a> clustering.</p>
New Tutorial on Principal Components Analysis2017-02-17T00:00:00+00:00http://uc-r.github.io/2017/02/17/pca<p><img src="/public/images/analytics/pca/unnamed-chunk-8-1.png" style="float:right; margin: 2px 0px 0px 10px; width: 40%; height: 40%;" /></p>
<p>Principal components analysis (PCA) reduces the dimensionality of our data, allowing most of the variability to be explained using fewer variables than the original data set. This allows us to understand the primary features that can best represent our data. Check out the latest <a href="http://uc-r.github.io/pca">tutorial</a> which covers PCA</p>
New Tutorial on Logistic Regression2017-02-03T00:00:00+00:00http://uc-r.github.io/2017/02/03/logistic-regression<p><img src="/public/images/analytics/logistic_regression/plot2-1.png" style="float:right; margin: 2px 0px 0px 10px; width: 40%; height: 40%;" /></p>
<p>Logistic regression is a foundational analytic technique for classification problems. It allows us to estimate the probability of a categorical response based on one or more predictor variables and tells us if the presence of a predictor increases (or decreases) the probability of a given outcome by a specific percentage. Check out the recently added <a href="http://uc-r.github.io/logistic_regression">tutorial on logistic regression</a>.</p>
New Tutorial on Linear Regression2017-01-20T00:00:00+00:00http://uc-r.github.io/2017/01/20/linear-regression<p><img src="/public/images/analytics/regression/sq.errors-1.png" style="float:right; margin: 2px 0px 0px 10px; width: 40%; height: 40%;" /></p>
<p>Linear regression is a useful and widely used statistical learning method and serves as a good jumping-off point for newer predictive analytic approaches. Check out the newly added <a href="http://uc-r.github.io/linear_regression">linear regression tutorial</a> that covers the basics of this power analytic technique.</p>
New Tutorials for Text Mining2017-01-07T00:00:00+00:00http://uc-r.github.io/2017/01/07/text-mining-tutorials<p>Analysts are often trained to handle tabular or rectangular data that are mostly numeric, but much of the data proliferating today is unstructured and typically text-heavy. Many of us who work in analytic fields are not trained in even simple interpretation of natural language. Fortunately, many of the principles used to organize, analyze, and visualize tabular data can be applied to unstructured text to extract meaning from this information.</p>
<p>Check out the first couple <a href="http://uc-r.github.io/descriptive">text mining tutorials</a> that have been released. Additional tutorials will be released in the coming weeks.</p>
New Tutorials for Better Workflow2016-09-17T00:00:00+00:00http://uc-r.github.io/2016/09/17/analytic-workflow<p>A.A. Milne said that <em>“organization is what you do before you do something, so that when you do it, it is not all mixed up.”</em> If you are not careful your data analyses can become an explosion of data files, R scripts, ggplot graphs, and final reports. Each project evolves and mutates in its own way and keeping all the files associated with a project organized together is a wise practice. Furthermore, reproducibility is crucial in any analytic project.</p>
<p>RStudio provides several great options for improving your analytic workflow. This includes RStudio projects, R Markdown, and the newly released R Notebook. Check out the tutorials we just released on these tools and get your analytic workflow moving in the right direction before your next project!</p>
<ul>
<li><a href="http://uc-r.github.io/r_projects">RStudio Projects</a></li>
<li><a href="http://uc-r.github.io/r_markdown">R Markdown</a></li>
<li><a href="http://uc-r.github.io/r_notebook">R Notebooks</a></li>
</ul>
New Tutorial on Creating Lollipop Charts2016-08-17T00:00:00+00:00http://uc-r.github.io/2016/08/17/lollipop-charts<p><img src="/public/images/visual/lollipop/lollipop-icon.png" style="display: block; margin: auto;" /></p>
<p>Check out the latest tutorial on creating lollipop charts <a href="http://uc-r.github.io/lollipop">here</a>. Enjoy!</p>
Updated Intro to R Bootcamp Files!2016-08-10T00:00:00+00:00http://uc-r.github.io/2016/08/10/updated-bootcamp-files<p>UC BANA students, the <a href="http://uc-r.github.io/r_bootcamp">Intro to R Bootcamp</a> slides have been updated to reveal the solutions for each <em>Your Turn!</em> exercise. Just scroll down to the <a href="http://uc-r.github.io/r_bootcamp#schedule">Schedule section</a> and click on the <i class="fa fa-file-powerpoint-o" aria-hidden="true"></i> icon for each session in and you’ll see the updated slides. I also add my code to the script files (<i class="fa fa-file-code-o" aria-hidden="true"></i>) so you can see my approach to many of the topics we covered. Enjoy!</p>
Welcome!2016-06-09T00:00:00+00:00http://uc-r.github.io/2016/06/09/welcome<p>Welcome to R programming at UC! This is a a new online repository hosted by University of Cincinnati’s Department of Operations, Business Analytics and Information Systems that provides tutorials to get you proficient in R programming. You will find a wide range of tutorials - from the basics of getting R and RStudio up and running to more advanced subjects such as regular expressions, analytic modeling, and visualization techniques. And we will continue to add more tutorials over time so follow us on Twitter at <a href="https://twitter.com/UC_Rstats">@UC-Rstats</a> so you can keep up to date on our most current offerings!</p>