This is an old revision of the document!
NOTES ABOUT MACHINE LEARNING AND ARTIFICIAL INTELLIGENCE AI
Notes:
Vectors and matrices are basic for machine learning.
BOOK Oreilly: 'Applied Machine Learning and AI for Engineers'
Jeff Proise github «
book
/Documents/PLURALSIGHT/datascience/Applied-Machine-Learning-main
source /Users/santosj/Documents/PLURALSIGHT/datascience/bin/activate
jupyter notebook /Users/santosj/Documents/PLURALSIGHT/datascience/Applied-Machine-Learning-main
-
Future reading: Machine Learning for Network and Cloud Engineers
External Link
Oreilly: Machine Learning with scikit-learn David Mertz
github
-
-
AI HARDWARE - GPUs
AMD Instinct MI series
Amazon's Inferentia (for machine learning inference on AWS)
Google's TPUs (Tensor Processing Units, custom hardware for Google’s machine learning tasks)
Intel Gaudi (designed for deep learning training)
NVIDIA GPUs (e.g., A100, H100, used for training and inference in deep learning applications)
NVIDIA Tensor Cores (hardware feature within NVIDIA GPUs, optimized for mixed-precision AI workloads)
Current practical models:
48
GB Nvidia RTX 6000 Ada graphics card
PRACTIAL NOTES ON MODELS:
Models multiply matrices.
Those matrices are multi-dimensionals : tensors
They are made of weight and bias « When defining a model weight and bias are called, generically, parameters.
Eg: 100B (all tensor's bias and weights, added together)
HF transformers library is ~different from transformers architecture. HF's is framework for loading, training, fine-tuning, and deploying transformer models across NLP and vision tasks. It provides access to thousands of pretrained models, simplifies workflows with task-specific pipelines, and supports custom training on new datasets. Beyond downloading models, Transformers enables production-ready deployment with optimizations for diverse hardware
HUGGINGFACE
models, datasets and prototypes
open-source and open-weight
we can download pre-trained Llama, via ollama and then fine-tune it.
One of the reason is so it identifies patterns bettwer (tex, images…). This process is called
embedding (Embeddings capture the inherent properties and relationships of the original data in a condensed format and are often used in Machine Learning use cases. See
Link « Better
classification