aiontext / output.json
Kushala's picture
Upload output.json
b556a80 verified
[
{
"output": "AION User Guide\n\n(Release 2.8.5)\n\nCopyright Notice\n\nThe information contained in this document is the property of HCL Technologies Limited. Except as specifically authorized in writing by HCL Technologies Limited, the holder of this document shall: (1) keep all information contained herein confidential and shall protect same in whole or in part from disclosure and dissemination to all third parties and, (2) use same for operating and maintenance purposes only. \n\n\u00a9Copyright 2018-2023 HCL Technologies Ltd \u2013 All rights reserved.\n\n\n\n\n\n\nIntroduction\n\n\n\nAION (Artificial Intelligence ON) is an AI life cycle management platform for applying machine learning to real-world problems. AION encompasses the complete pipeline from raw data ingestion to a deployable machine learning model with a less-code/no-code approach. It includes the following sub-processes:\n\nAutoML- Enable business user-generated ML models with no code approach\n\nMLOPs - Model Versioning, Deployment & Drift Monitoring\n\nMLaC \u2013 Machine Learning as Code for automatic pipeline code generation\n\nExplainability- Higher accuracy predictions with supported explanations & model confidence \n\nIn AION, all these processes are automated. When you give input data and the use case that needs to be solved, AION runs through all the stages of data analytics cycle and generates the best deployment model.\n\nAION Engines\n\nAION has a multi-stage pipeline as depicted below in the figure. A brief explanation of each stage is given below:\n\nINGESTOR- Data Ingestion: In this stage, the dataset is uploaded in AION GUI from disparate sources.\n\nEXPLORER \u2013 Exploratory Data Analysis: This shows the details about the nature of data, the relation between features, model statistics, and performance information to derive descriptive insights.\n\nTRANSFORMER- Data Processing: In this stage data-cleaning, data preparation, and outlier detectionares did automatically to improve the quality of data for better model accuracy.",
"instruction ": "",
"input": ""
},
{
"output": "SELECTOR- Feature Selection: The statistical analysis takes place to identify the relevant features for model training and remove the unimportant features based on correlation & importance.\n\nLEARNER- Model Training Hyper Parameter Tuning: This stage trains configured models and selects the best parameters based on hyperparameter tuning (using which the models are trained). A broad spectrum of algorithms is supported.\n\nPREDICTOR- Inference service: ML Model serving & inference services.\n\nOBSERVER \u2013 Model Monitoring: It monitors the model for input & output drift of data or predictions. \n\nEXPLAINER- Explainable AI: It explains the model & Uncertainty quantification of the prediction.\n\nCONVERTOR- Model Conversion for Edge Device: The generated models can be converted to any desired format that is suitable for embedded, cloud or edge deployments. \n\nTESTER- Model Testing: Once the models are generated, it can be tested using different testing methodologies.\n\n\n\nCODER- Machine Learning as Code: It creates Python code automatically for ML pipeline components.\n\nLaunching AION\n\nAION Explorer is the application used to launch the AION GUI. Followings steps are followed to lunch the GUI.\n\nDouble click on Explorer icon on desktop to launches the AION GUI.\n\n AION explorer shell will pop up loading all the requirements.\n\nAfter a while GUI landing page will appear in a browser as shown in the figure below. Users need to reload in case the GUI main page does not appear.\n\n\n\nNote:\n\n Access to Hyperscale like AWS, GCP, Azure and data sources like graviton and NiFi will not be possible on HCL network.\n\nTo enable access user needs to raise service exchange ticket and raise access to this source.\n\nCompatibility with previous version:",
"instruction ": "",
"input": ""
},
{
"output": "Compatibility with previous version:\n\nNote: Note: if a previous version of AION is already installed, re-training of use cases of that version is possible in the latest version, and it is backward compatible. But, use cases created in the latest version may not run the previous version, as forward compatibility is not supported in all cases\n\nModel Management\n\nCreate Usecase\n\nTo create a new usecase, follow the steps given below:\n\nIn the left panel, click Home. \n\nClick Add Use Case tab, enter the Use case name or keep it default and its Description. \n\nNote: Use case name does not support any special characters or space. \n\nClick Submit.\n\nAs a result, the usecase is added in the USE CASE LIST. All the usecases listed in the USE CASE LIST can be viewed as shown in the figure below. \n\nCompute Infrastructure\n\nUsers can select following computer infrastructure for training:\n\nLocal\n\nAWS EC2 (AION)\n\nAWS EC2 (LLaMA-7B)\n\nGCP GCE (LLaMA-13B)\n\nClick on the Compute Infrastructure tab as shown below. \n\nAWS EC2 (AION): User can use the Amazon AWS EC2 infrastructure with default settings or can configure the details from settings tab as shown below.\n\n\n\nAWS EC2 (LLaMA-7B) and GCP GCE (LLaMA-13B) for LLM fine tuning.\n\nView Existing Models\n\nTo view the existing models, follow the steps given below:\n\nIn the left panel, click Home.\n\nThe existing use cases and their training status can be seen.\n\nTrained models can be lunched for Retaining, Prediction, Trusted AI parameters and Drift Analysis. \n\nFor multiple versions of use case, click on use case name to view different model versions that are trained by Model Retraining. Eg: By clicking on AI00053_lanquos, different versions under that Use Case can be viewed as shown in the figure below.\n\nPublish Model \n\nTrained model can be published in the form of following packages: \n\nMLaC\n\nFederated Learning\n\nPython Package\n\nHomomorphic Encryption\n\nModel Container\n\nRetrained Model",
"instruction ": "",
"input": ""
},
{
"output": "Model Container\n\nRetrained Model\n\nMLaC: Machine Learning as Code generates code automatically based on ML Operations performed during various ML Lifecycle. Using MLaC, expert data scientists can have better control over experimentation, optimization & deployment of ML Models.\n\nMLaC has four unique components. \n\nGenerate Code: Generates the code for each pipeline component of the selected use case i.e., data transformations, feature engineering, model training and prediction. This code can further be used to retrain the model on the latest data and can also be used to create a container for each component. The user will find the code in the local system through path: C:\\Users\\yashaswini.ragi\\AppData\\Local\\HCLT\\AION\\target\\AION_375_1\\publish\\MLaC\n\nGenerate Container (Local):\n\nPre-requisite:\n\n Local Machine should have docker installed.\n\nCreate docker images for each of the components by running the docker file present in each component.\n\n Once the docker images are ready to consume, create the container from the images and run that container sequentially as mentioned below:\n\nModel Monitoring > Data Ingestion > Data Transformation > Feature Engineering > Model Training > Model Registry > Model Serving\n\nUser can see the status of each step through the shell as shown below.\n\nExec Pipeline (Local):\u00a0Another way to run this component without using docker is to run the command run_pipeline.py in aion shell.\n\nThe command will execute each component one by one locally in the user system. \n\n\n\nExec Kubeflow pipeline: Using Kubeflow user can run all the components as a single pipeline, which is managed by Kubeflow internally.\n\nGitHub Upload:\u00a0Upload the Generated Code to the GitHub repository.\n\nMLaC service can be used for prediction through PostMan following the steps given below-\n\nOpen PostMan desktop app.\n\nConfigure a POST request with the predict API of the form http://<localhost:port>//AION//<usecasename>/predict \n\nEg: http://127.0.0.1:8094//AION/AI0192/predict.\n\nSelect authorization as Basic Authorization",
"instruction ": "",
"input": ""
},
{
"output": "Select authorization as Basic Authorization\n\nUsername: USERNAME\n\nPassword: PASSWORD\n\nPut the data in the Body as a raw json format. \n\nClick send to perform prediction.\n\nNote: For more information users can follow the user guide for MLaC by clicking on the icon on the top right corner of the pop-up window.\n\n\n\nFederated Learning (Beta): Federated learning is a way to train AI models without exposing the data, offering a way to unlock information to feed new AI applications. Federated learning (or collaborative learning) is an approach used to train a decentralized machine learning model across multiple partitions. The federated learning participants can be edge devices, data centers, smartphones, medical wearables, IoT devices, vehicles, etc. They collaboratively train a shared model while keeping the training data locally without exchanging it. Each participant downloads the model from the server. They train it on their private data, then summarize and encrypt the model\u2019s new configuration. The model updates are sent back to the cloud, decrypted, averaged, and integrated into the centralized model. Then, this centralized model is once again shared with participants for the next round of federated learning. Iteration after iteration, the collaborative training continues until the model is fully trained. In AION, Federated Learning is supported for the following algorithms: Logistic Regression, Linear Regression, and Simple Neural Network (SNN). \n\nPython Package: Python Package creates a WHL File of a trained model so that the user can further use that file in testing and prediction. If the user wants to consume the trained model for prediction in some system where AION is not there, then either the user can install a python package or use the docker images.\n\nPython package can be consumed by following the steps:\n\nClick on the python package icon from the packages tab and as a result whl file will get download.\n\nOpen the command prompt and type the command: python -m pip install whl_file_path and press Enter.",
"instruction ": "",
"input": ""
},
{
"output": "Model container: The model container creates a docker image of the trained model so that user can use that image anywhere, pull the image, and perform testing. If the user wants to consume the trained model for prediction purpose in some system where AION is not there, then either the user can install it as python package or use the docker images. \n\nThere are two cased to create a docker container:\n\nWhen the Docker is installed in the machine user can click on the docker and it gets created. \n\nIn case Docker is not installed, running aion will throw an error. For that user can copy the mentioned path file and put in other machine where the docker should be installed and can build the image there also using docker build command.\n\nThe model container also supports CORS (Cross-Origin Resource Sharing) protocol to mitigate the default security policy, Same-Origin Policy (SOP) used by the browsers to protect resources. The same-origin policy is a critical security mechanism that restricts a documents or script loaded by one origin can interact with a resource from another origin.\n\n\n\nCommand to enable CORS: python aion_service.py ip 0.0.0.0 -p 80 -cors \u201chttp://localhost:5000\u201d\n\nHomomorphic Encryption: Homomorphic Encryption (HE) combines data encryption with privacy protection. A homomorphic cryptosystem is like other forms of public encryption in that it uses a public key to encrypt data and allows only the individual with the matching private key to access its unencrypted data. However, what sets it apart from other forms of encryption is that it uses an algebraic system to allow a variety of computations on the encrypted data. AION Supports only the XGBoost algorithm.\n\nTo publish the trained model, go to the Packages tab and click on the icons of above-mentioned packages accordingly. \n\nUsers can select any model or any version of that model for publishing so that it can be installed on the system for faster inference.",
"instruction ": "",
"input": ""
},
{
"output": "Eg: If an user clicks on Python Package icon from the Packages tab, a pop-up window will appear to confirm the model download. Click Ok and then the model gets downloaded in the form of a WHL file which can be used for further testing and prediction.\n\nModel Re-training\u00a0\n\nAfter the Model got trained go to the Home.\n\nFrom the Train column click on the Model Re-training icon to re-train the model withing Re-Configuring.\n\nFor Re-Configuring and training click on the icon.\n\nClick On EDA if required or click on Next. \n\nSet the Basic and Advance configuration if required.\n\nClick Train Model tab to train the model.\n\n\n\nModel will get Re-trained successfully and user can see the results as shown below.\n\nSystem Setting\n\nTo configure and change system settings, \n\nIn the left panel, click the Settings icon. \n\nUser can view the settings and configure them if required.\n\nConfigurations for the following integration entities are available:\n\nConfiguration Settings: \n\nApache Kafka Configuration: Kafka cluster by default will be running in the local host. Kafka uses key-value pairs in the\u00a0property file format\u00a0for configuration. The default values of Kafka cluster ip and kafka cluster port are localhost and 6000 respectively. The value of Kafka Topic can be supplied by the user as per the Kafka configuration.\n\nOpen AI Configuration: For open AI related task Open API key, Open API Base, Open API Type, and Open API Version need to be given by the user.\n\n\n\nCloud Storage Settings:\n\nAWS S3 Bucket\n\nGCP Bucket\n\nAzure Storage\n\nCredentials need to be given for accessing AWS, GCP and Azure Storage by clinking on add bucket and data hence data can be uploaded from the cloud in GUI.\n\nCloud Infrastructure Settings: The user can configure the details of AWS EC2 compute infrastructure to consume AWS EC2 infrastructure.\n\nModel Training\n\n\n\nThe Model Training tab is used to train the model based on the data input file.\n\nData Upload \n\nData upload tab is used to load the dataset in the GUI to train the model and get insight of it.",
"instruction ": "",
"input": ""
},
{
"output": "While uploading the dataset, data size and ram size need to be checked in two cases:\n\nIf file size > ram size, file will not be uploaded for further process. File size should not be more than 80% of ram size.\n\nIf the file size > 50% of ram size than the alert will be shown and proceed for the next step. \n\nStructured Data\n\nTo upload the data file, follow the steps given below:\n\nCreate use case.\n\nClick the Data Upload tab.\n\nFile like csv, tar and zip file can be uploaded from local, URL, NiFi, GCP Storage, Azure Storage, AWS S3 Bucket, and Graviton. Credentials might be required to ingest data form all the cloud storages. \n\nAs we Submit, the file will be uploaded successfully.\n\nGraviton: It enables implementation of the modern data platform using polyglot data technologies, open-source data frameworks and cloud-native services. It facilitates the development and delivery of optimal data products and efficient management of data product lifecycle from development to production.\n\nUnstructured Data\n\nThis option is used to upload a document.\n\nTo upload a file, follow the steps given below:\n\nClick on the Data Upload tab.\n\nClick Unstructure data.\n\nSelect the type of document you want to upload (txt, log, pdf).\n\nPlace the files in a folder and specify the path in the Folder Path field.\n\n\n\nClick Submit.\n\nAs a result, the files in the folder will be uploaded successfully.\n\nCustom Python Script\n\nThe python script needs to be created by the user, which can be converted into a data file.\n\nTo upload a python script, follow the steps given below:\n\nClick the Data Upload tab.\n\nClick Custom Python Script.\n\nSpecify the python script path.\n\nClick Submit.\n\nAs a result, the file will be uploaded successfully.\n\nExternal Data Sources\n\nUsers can also access data from external databases for model training.\n\nFor external systems follow the following steps:\n\nClick the Data Upload tab.\n\nClick External Data Sources.",
"instruction ": "",
"input": ""
},
{
"output": "Click External Data Sources.\n\nSelect different sql databases and data warehouses to ingest data. viz.- SQLite, Microsoft SQL, PostgreSQL, MySQL, Oracle, Google BigQuery, Redshift, Snowflake, Actian.\n\n\n\nTo access data from the database, you need to provide essential credentials like Database Name, Username, Host Port, etc.\n\nIf a single table contains all the data, you can simply put the table name and fetch the table as Data Frame. Otherwise, the user can perform join and selection based on the condition to merge more than one table and finally fetch the data.\u00a0\n\nNote: As of now, this option is to upload data from some external databases and some data warehouses like- SQLite, PostgreSQL, etc. Users can insert a single table as a database. Also, some basic SQL operation like join, and selection based on where is also possible on multiple tables. Please contact ERS Research in case you wish to enable this option.\n\nSKLearn Data Sources\n\nTo upload a standard sklearn data set, follow the steps given below:\n\nClick on the Data Upload tab.\n\nClick SKLearn Data Sources.\n\nSelect a dataset.\n\nClick Submit.\n\nAs a result, the dataset will be uploaded successfully.\n\nAION Labelled Data\n\nHere the trained unsupervised model can be used as a base model to train the supervised model. The two main significance of using a model within model is to simplify a large, complex model and allow for more advanced use of model\u00a0iterators.\n\nTo use the model chain following steps are followed:\n\n\n\nSubmit new use case \uf0e0 Upload dataset\n\nTrain unsupervised model i.e., clustering \u00a0\n\nCreate a new use case \uf0e0 upload the dataset. \n\nTraining a supervised model i.e., classification, regression, etc.\n\nFrom the drop-down list choose the model as the base model to train the new supervised model.\n\nClick next. \n\nEDA\n\nEDA stands for Exploratory Data Analysis where user can analyze and investigate data sets and summarize their main characteristics, employing data visualization methods.\n\nAfter uploading the data file, you can view the dataset of top 10 records.",
"instruction ": "",
"input": ""
},
{
"output": "To see the remaining details, click the EDA button.\n\nOn clicking the EDA button, a window will pop up that enables configuring the features and the sample of data that needed to be selected for EDA.\n\nSelect features and data subsampling size for exploratory data analysis. Finally, click Next.\n\n\n\nData Overview: The summary statistics of the features can be viewed as shown in the figure below.\n\nData Distribution: It shows the data distribution type for each feature of a dataset. \n\nFor numerical data, select the numerical feature and click on Show Data Distribution to see the distribution of data is generated as shown below.\n\nFor text data, select the text feature and click on the Show Data Distribution. Hence, word cloud is generated as shown below. \n\n\n\nMore details can be seen by clicking the icon on the right of the Data Distribution tab. \n\nTimeseries Analysis: Check the dataset Stationarity and Seasonality for each of the feature of the ingested dataset. \n\nFeature Importance: The relevant features depending on the variance ratio can be viewed in the bar chart. The higher the variance ratio more relevant the feature is. For example: From the figure below, the sepal_width is the most significant feature for iris data classification.\n\nCorrelation Analysis: The strength of relationships varying from 0 to 1 among various features can be viewed as shown in the figure below. A value closer to 1 shows that the features are highly correlated and a value close to 0 shows that the features are less correlated. Example: The correlation value of feature row record_dateTime and feature column id is 1 which shows that these two features are highly correlated and considering both the feature is irrelevant. \n\n\n\nUnsupervised Clustering: The unsupervised clustering of the data can be viewed as shown in the figure. Example: From the figure, we observed five clusters are being created and the average value of those clusters for feature retransmission is 5.52, 3.98, 5.23, 4.66 & 5.44.",
"instruction ": "",
"input": ""
}
]