Foundations
The Foundations module comprises two courses where we get our hands dirty with Python programming language for Artificial Intelligence and Machine Learning and Statistical Learning, head-on. These two courses set our foundations for Artificial Intelligence and Machine Learning online course so that we sail through the rest of the journey with minimal hindrance. Welcome to the program.
Self-paced Module: Introduction to Data Science and AI
Gain an understanding of the evolution of AI and Data Science over time, their application in industries, the mathematics and statistics behind them, and an overview of the life cycle of building data driven solution.
- The fascinating history of Data Science and AI
- Transforming Industries through Data Science and AI
- The Math and Stats underlying the technology
- Navigating the Data Science and AI Lifecycle
Self-Paced Module: Python Pre-Work
This course provides you with a fundamental understanding of the basics of Python programming and builds a strong foundation of the basics of coding to build AI and Machine Learning (ML) applications
- Introduction to Python Programming
- AI Application Case Study
Module 1: Python Foundations
1 Project
1 Project
Python is an essential programming language in the tool-kit of an AI & ML professional. In this course, you will learn the essentials of Python and its packages for data analysis and computing, including NumPy, SciPy, Pandas, Seaborn and Matplotlib.
- Python Programming Fundamentals
Python is a widely used high-level, interpreted programming language, having a simple, easy-to-learn syntax that highlights code readability.
This module will teach you how to work with Python syntax to executing your first code using essential Python fundamentals
- Python for Data Science - NumPy and Pandas
NumPy is a Python package for scientific computing like working with arrays, such as multidimensional array objects, derived objects (like masked arrays and matrices), etc. Pandas is a fast, powerful, flexible, and simple-to-use open-source library in Python to analyse and manipulate data.
This module will give you a deep understanding of exploring data sets using Pandas and NumPy.
- Exploratory Data Analysis
Exploratory Data Analysis, or EDA, is essentially a type of storytelling for statisticians. It allows us to uncover patterns and insights, often with visual methods, within data.
This module will give you a deep insight into EDA in Python and visualization tools-Matplotlib and Seaborn.
Data preprocessing is a crucial step in any machine learning project and involves cleaning, transforming, and organizing raw data to improve its quality and usability. The preprocessed data is used both analysis and modeling.
Text data is one of the most common forms of data and analyzing it plays a crucial role in extracting valuable insights from unstructured information in human language. This module covers different text processing and vectorization techniques to efficiently extract information from raw textual data.
Self-paced Module: Statistical Learning
3 Quizzes
1 Project
3 Quizzes
1 Project
Statistical Learning is a branch of applied statistics that deals with Machine Learning, emphasizing statistical models and assessment of uncertainty. This course on statistics will work as a foundation for Artificial Intelligence and Machine Learning concepts learnt in this AI ML PG program.
- Descriptive Statistics
The study of data analysis by describing and summarising numerous data sets is called Descriptive Analysis. It can either be a sample of a region’s population or the marks achieved by 50 students.
This module will help you understand Descriptive Statistics in Python for AI ML.
- Inferential Statistics
Inferential Statistics helps you how to use data for estimation and assess theories. You will know how to work with Inferential Statistics using Python.
- Probability & Conditional Probability
Probability is a mathematical tool used to study randomness, like the possibility of an event occurring in a random experiment. Conditional Probability is the likelihood of an event occurring provided that several other events have also occurred.
In this module, you will learn about Probability and Conditional Probability in Python for AI ML.
- Hypothesis Testing
Hypothesis Testing is a necessary Statistical Learning procedure for doing experiments based on the observed/surveyed data.
You will learn Hypothesis Testing used for AI and ML in this module.
- Chi-square & ANOVA
Chi-Square is a Hypothesis testing method used in Statistics, where you can measure how a model compares to actual observed/surveyed data.
Analysis of Variance, also known as ANOVA, is a statistical technique used in AI and ML. You can split observed variance data into numerous components for additional analysis and tests using ANOVA.
This module will teach you how to identify the significant differences between the means of two or more groups.
Machine Learning
The next module is the Machine Learning online course, where you will learn Machine Learning techniques and all the algorithms popularly used in Classical ML that fall in each category.
Module 2: Machine Learning
1 Project
1 Project
In this module, understand the concept of learning from data, build linear and non-linear models to capture the relationships between attributes and a known outcome, and discover patterns and segment data with no labels.
Supervised Machine Learning aims to build a model that makes predictions based on evidence in the presence of uncertainty. In this course, you will learn about Supervised Learning algorithms of Linear Regression and Logistic Regression.
Linear Regression is one of the most popular supervised ML algorithms used for predictive analysis, resulting in producing the best outcomes. You can use this technique to assume a linear relationship between the independent variable and the dependent variable. You will cover all the concepts of Linear Regression in this module.
A decision tree is a Supervised ML algorithm, which is used for both classification and regression problems. It is a hierarchical structure where internal nodes indicate the dataset features, branches represent the decision rules, and each leaf node indicates the result.
Unsupervised Learning finds hidden patterns or intrinsic structures in data. In this machine learning online course, you will learn about commonly-used clustering techniques like K-Means Clustering and Hierarchical Clustering along with Dimension Reduction techniques like Principal Component Analysis.
K-means clustering is a popular unsupervised ML algorithm, which is used for resolving the clustering problems in Machine Learning. In this module, you will learn how the algorithm works and later implement it. This module will teach you the working of the algorithm and its implementation.
Module 3: Advanced Machine Learning
1 Project
1 Project
Ensemble methods help to improve the predictive performance of Machine Learning models. In this machine learning online course, you will learn about different Ensemble methods that combine several Machine Learning techniques into one predictive model in order to decrease variance, bias or improve predictions.
- Bagging and Random Forests
In this module, you will learn Random Forest, a popular supervised ML algorithm that comprises several decision trees on the provided several subsets of datasets and calculates the average for enhancing the predictive accuracy of the dataset, and Bagging, an essential Ensemble Method.
Boosting is an Ensemble Method which can enhance the stability and accuracy of machine learning algorithms, converting them into robust classification, etc.
Model tuning is a crucial step in developing ML models and focuses on improving the performance of a model using different techniques like feature engineering, imbalance handling, regularization, and hyperparameter tuning to tweak the data and the model. This module covers the different techniques to tune the performance of an ML model to make it robust and generalized.
Artificial Intelligence & Deep Learning
The AI and Deep Learning course will take us beyond the traditional ML into the realm of Neural Networks. From the regular tabular data, we move on to training our models with unstructured data like Text and Images.
Module 4: Introduction to Neural Networks
1 Project
1 Project
In this module, implement neural networks to synthesize knowledge from data, demonstrate an understanding of different optimization algorithms and regularization techniques, and evaluate the factors that contribute to improving performance to build generalized and robust neural network models to solve business problems.
- Deep Learning and its history
Deep Learning carries out the Machine Learning process using an ‘Artificial Neural Net’, which is composed of several levels arranged in a hierarchy. It has a rich history that can be traced back to the 1940s, but significant advancements occurred in the 2000s with the introduction of deep neural networks and the availability of large datasets and computational power.
The multilayer perceptron (MLP) is a type of artificial neural network with multiple layers of interconnected neurons, including an input layer, one or more hidden layers, and an output layer. It is a versatile architecture capable of learning complex patterns from data.
Activation Function is used for defining the output of a neural network from numerous inputs.
Backpropagation is a key algorithm used in training artificial neural networks, enabling the calculation of gradients and the adjustment of weights and biases to iteratively improve the performance of a neural network.
Optimizers are algorithms used to adjust the parameters of a neural network model during training to minimize the loss function. Different types of optimizers are Gradient Descent, RMSProp, Adam, etc.
- Weight Initialization and Regularization
Weight initialization is the process of setting initial values for the weights of a neural network, which can significantly impact the model's training and convergence. Regularization is a technique used in machine learning/ neural networks to prevent the model from overfitting, which helps improve the model's generalization ability.
Module 5: Natural Language Processing with Generative AI
1 Project
1 Project
This course will help you get introduced to the world of natural language processing, gain a practical understanding of text embedding methods, gain a practical understanding of the working of different transformer architectures that lie at the core of large language models (LLMs), explore how retrieval augmented generation (RAG) integrates information retrieval to improve the accuracy and relevance of responses from an LLM, and design and implement robust NLP solutions using open-source LLMs combined with prompt engineering techniques.
Natural Language Processing (NLP) is a branch of AI that focuses on processing and understanding human language to facilitate the interaction of machines with it. Word embeddings allow us to numerically represent complex textual data, thereby enabling us to perform a variety of operations on them. This module introduces participants to the world of NLP, covers different word embedding techniques, and the steps involved in designing and implementing hands-on solutions combining word embedding methods with machine learning techniques for solving NLP problems
- Attention Mechanism and Transformers
Transformers are neural network architectures that develop a context-aware understanding of data and have revolutionized the field of NLP by exhibiting exceptional performance across a wide variety of tasks. This module dives into the underlying workings of different transformer architectures and how to use them to solve complex NLP tasks.
- Large Language Models and Prompt Engineering
Large Language Models (LLMs) are ML models that are pre-trained on large corpora of data and possess the ability to generate coherent and contextually relevant content. Prompt engineering is a process of iteratively deriving a specific set of instructions to help an LLM accomplish a specific task. This module introduces LLMs, explains their working, and covers practices to effectively devise prompts to solve problems using LLMs.
- Retrieval Augmented Generation
Retrieval augmented generation (RAG) combines the power of encoder and generative models to produce more informative and accurate outputs from a knowledge base. This module will provide a thorough coverage of leveraging sentence transformers to encode data, vector databases to store and efficiently retrieve information from the encoded data, and LLMs to use the information to enhance the quality and relevance of the generated output.
Module 6: Introduction to Computer Vision
1 Project
1 Project
This course will introduce you to the world of computer vision, demonstrate an understanding of image processing and different methods to extract informative features from images, build convolutional neural networks (CNNs) to unearth hidden patterns in image data, and leverage common CNN architectures to solve image classification problems.
Computer Vision is a branch of AI that focuses on understanding and extracting meaningful insights from image data. This module provides an overview of the world of computer vision and covers techniques to process images and extract meaningful patterns from them.
- Convolutional Neural Networks
Given the complex nature of image data, convolutional neural networks (CNNs) are utilized to capture relevant spatial information in images. Transfer learning is a method to leverage the underlying knowledge needed to solve one problem to other problems. This module will cover the fundamentals of CNNs, how to build them from scratch, and how to leverage common CNN architectures via transfer learning to solve different image classification problems
Module 7: Model Deployment
This course will help you comprehend the role of model deployment in realizing the value of an ML model and how to build and deploy an application using Python.
- Introduction to Model Deployment
Model deployment is the process of making a trained machine learning model accessible to a wider audience by operationalizing it. This module introduces participants to model deployment, provides an overview of its need in generating business value from ML models, and serializing and deploying ML models using Python libraries like Streamlit.
Containerization is the process of packaging applications and their dependencies into self-contained units called containers to ensure consistent execution across different environments. This module dives into packaging ML models and their dependencies into containers using Docker and simplifying deployment of the ML models using Python libraries like Flask.
Self-paced Module: Generative AI
Get an overview of Generative AI, what ChatGPT is and how it works. delve into the business applications of ChatGPT, and an overview of other generative AI models/tools via demonstrations.
- ChatGPT and Generative AI - Overview
- ChatGPT - Applications and Business
- Breaking Down ChatGPT
- Limitations and Beyond ChatGPT
- Generative AI Demonstrations
Self-paced Module: Recommendation Systems
1 Project
1 Project
The last module in this Artificial Intelligence and Machine Learning online course is Recommendation Systems. A large number of companies use recommender systems, which are software that select products to recommend to individual customers. In this course, you will learn how to produce successful recommender systems that use past product purchase and satisfaction data to make high-quality personalized recommendations.
- Popularity-based Model
A popularity-based model is a recommendation system, which operates based on popularity or any currently trending models.
- Market Basket Analysis
Market Basket Analysis, also called Affinity Analysis, is a modeling technique based on the theory that if you purchase a specific group of items, then you are more probable to buy another group of items.
- Content-based Model
First, we accumulate the data explicitly or implicitly from the user. Next, we create a user profile dependent on this data, which is later used for user suggestions. The user gives us more information or takes more recommendation-based actions, which subsequently enhances the accuracy of the system. This technique is called a Content-based Recommendation System.
- Collaborative Filtering
Collaborative Filtering is a collective usage of algorithms where there are numerous strategies for identifying similar users or items to suggest the best recommendations.
- Hybrid Recommendation Systems
A Hybrid Recommendation system is a combination of numerous classification models and clustering techniques. This module will lecture you on how to work with a Hybrid Recommendation system.
Self-paced Module: Multimodal Generative AI
This course will help you gain an understanding of the core concepts of databases and SQL, gain practical experience writing simple SQL queries to filter, manipulate, and retrieve data from relational databases, and utilize complex SQL queries with joins, window functions, and subqueries for data extraction and manipulation to solve real-world data problems and extract actionable business insights.
- Introduction to DB and SQL
- Fetching, Filtering, and Aggregating Data
- Inbuilt and Window Functions
- Joins and Subqueries
Self-paced Module: Introduction to SQL
This course will help you gain an understanding of the core concepts of databases and SQL, gain practical experience writing simple SQL queries to filter, manipulate, and retrieve data from relational databases, and utilize complex SQL queries with joins, window functions, and subqueries for data extraction and manipulation to solve real-world data problems and extract actionable business insights.
- Introduction to DB and SQL
- Fetching, Filtering, and Aggregating Data
- Inbuilt and Window Functions
- Joins and Subqueries
Career Assistance: Resume and LinkedIn profile review, interview preparation, 1:1 career coaching
This post-graduate certification program on artificial intelligence and machine learning will assist you through your career path to building your professional resume and reviewing your Linkedin profile. The program will also conduct mock interviews to boost your confidence and nurture you nailing your professional interviews. The program will also assist you with one-on-one career coaching with industry experts and guide you through a career fair.
Post Graduate Certificate from The University of Texas at Austin
Earn a Postgraduate Certificate in the top-rated Artificial Intelligence and Machine Learning online course from the University of Texas, Austin. The course’s comprehensive Curriculum will foster you into a highly-skilled professional in Artificial Intelligence and Machine Learning. It will help you land a job at the world’s leading corporation and power ahead your career transition.
On-Campus Immersion in Decision Science and AI (Optional Paid Program)
The Decision Science and AI is a 3-day on-campus Program that presents a valuable opportunity to explore AI use cases and become a driving force behind AI-driven initiatives within your organization. It comprises of dynamic discussions, collaboration with like-minded professionals, and engaging networking sessions hosted at the prestigious University of Texas at Austin.
- Welcome & Program Orientation
- Introduction to Decision Sciences & AI
- Campus Tour & Group Photo
- Introduction to Dynamic Programming
- Programming an AI agent to Play a Variant of Blackjack
- Introduction to Reinforcement Learning
- Programming an AI Agent that learns by itself to play computer games
- Session with Industry Mentor
- The Art and Science of Negotiations
- Project Brief and Active group work
- Group work on Project
- Certifications and Photo Ops