Lightwood
Lightwood is the default AI engine used in MindsDB. It deals mainly with classification, regression, and time-series problems in machine learning.
By providing it with the input data and problem definition, Lightwood generates predictions following three core steps that include Data pre-processing and cleaning, Feature engineering, and Model building and training. The input data ranges from numbers, dates, categories, text, quantities, arrays, matrices, up to images, audios, and videos (passed as URLs).
How It Works
Here is the algorithm followed by Lightwood starting from the input data setup, through model building and training, up to getting predictions.
The input data is pre-processed and each column is assigned a data type. Next, data is converted into features via encoders that transform data into numerical representation used by the model. Finally, a predictive model takes the encoded feature data and outputs a prediction for the target.
Under the hood, the model splits data into the training, validation, and testing sets, with ratios that are dynamic but usually an 80-10-10 ratio. The split is done by default using random sampling without replacement, stratified on the target column. Doing so, it determines the accuracy of the model by evaluating on the held out test set.
Users can either use Lightwood’s default mixers/models or create their own approaches inherited from the BaseMixer
class.
To learn more about Lightwood philosophy, follow this link.
Accuracy Metrics
Lightwood provides ways to score the accuracy of the model using one of the accuracy functions.
The accuracy functions include mean_absolute_error
, mean_squared_error
, precision_score
, recall_score
, and f1_score
.
You can define the accuracy function of choice in the USING
clause of the CREATE MODEL
statement.
Here are the accuracy functions used by default:
- the
r2_score
value for regression predictions. - the
balanced_accuracy_score
value for classification predictions. - the
complementary_smape_array_accuracy
value for time series predictions.
The values vary between 0 and 1, where 1 indicates a perfect predictor, based on results obtained for a held-out portion of data (i.e. testing set).
You can check accuracy values for models using the DESCRIBE
statement.
Tuning the Lightwood ML Engine Features
Description
In MindsDB, the underlying AutoML models are based on the Lightwood engine by default. This library generates models automatically based on the data and declarative problem definition. But the default configuration can be overridden using the USING
statement that provides an option to configure specific parameters of the training process.
In the upcoming version of MindsDB, it will be possible to choose from more ML frameworks. Please note that the Lightwood engine is used by default.
Syntax
Here is the syntax:
encoders
Key
It grants access to configure how each column is encoded. By default, the AutoML engine tries to get the best match for the data.
To learn more about encoders
and their options, visit the Lightwood documentation page on encoders.
model
Key
It allows you to specify the type of machine learning algorithm to learn from the encoder data.
Here are the model options:
Model | Description | |
---|---|---|
BaseMixer | It is a base class for all mixers. | |
LightGBM | This mixer configures and uses LightGBM for regression or classification tasks depending on the problem definition. | |
LightGBMArray | This mixer consists of several LightGBM mixers in regression mode aimed at time series forecasting tasks. | |
NHitsMixer | This mixer is a wrapper around an MQN-HITS deep learning model. | |
Neural | This mixer trains a fully connected dense network from concatenated encoded outputs of each feature in the dataset to predict the encoded output. | |
NeuralTs | This mixer inherits from Neural mixer and should be used for time series forecasts. | |
ProphetMixer | This mixer is a wrapper around the popular time series library Prophet. | |
RandomForest | This mixer supports both regression and classification tasks. It inherits from sklearn.ensemble.RandomForestRegressor and sklearn.ensemble.RandomForestClassifier. | |
Regression | This mixer inherits from scikit-learn’s Ridge class. | |
SkTime | This mixer is a wrapper around the popular time series library sktime. | |
Unit | This is a special mixer that passes along whatever prediction is made by the target encoder without modifications. It is used for single-column predictive scenarios that may involve complex and/or expensive encoders (e.g. free-form text classification with transformers). | |
XGBoostMixer | This mixer is a good all-rounder, due to the generally great performance of tree-based ML algorithms for supervised learning tasks with tabular data. |
Please note that not all mixers are available in our cloud environment. In particular, LightGBM, LightGBMArray, NHITS, and Prophet.
To learn more about all the model
options, visit the Lightwood documentation page on mixers.
Other Keys Supported by Lightwood in JsonAI
The most common use cases of configuring predictors use encoders
and model
keys explained above. To see all the available keys, check out the Lightwood documentation page on JsonAI.
Example
Here we use the home_rentals
dataset and specify particular encoders
for some columns and a LightGBM model
.
You can visit the comprehensive Lightwood docs here.
Check out the Lightwood tutorials here.