machine-learning - Concepts
Explore concepts tagged with "machine-learning"
Total concepts: 74
Concepts
- Supervised Learning - A machine learning approach where models are trained on labeled data with known correct outputs.
- AI Frontier Model - The most capable and advanced AI models at the cutting edge of performance, typically from leading AI labs.
- Neural Architecture Search (NAS) - Automated process of discovering optimal neural network architectures using machine learning rather than manual design.
- Computer Vision - A field of AI that enables computers to interpret and understand visual information from the world, including images and video.
- Multi-Task Learning - A machine learning approach where a single model is trained on multiple related tasks simultaneously, leveraging shared representations to improve generalization.
- Sparse Models - Neural network architectures where only a fraction of parameters are activated for any given input, enabling larger model capacity with lower computational cost.
- Big Data - Datasets so large, fast-moving, or complex that traditional data processing methods cannot handle them effectively, characterized by volume, velocity, variety, veracity, and value.
- Artificial Intelligence - The field of computer science focused on creating systems that can perform tasks requiring human-like intelligence, learning, and reasoning.
- Neural Networks - Computing systems inspired by biological neural networks in the brain, designed to recognize patterns and learn from data.
- Semantic Search - A search technique that finds information based on meaning and intent rather than exact keyword matching.
- Model Parameters - The learned numerical values (weights and biases) within a neural network that determine how the model transforms inputs into outputs.
- Ensemble Learning - A machine learning paradigm that combines predictions from multiple models to produce more accurate and robust results than any single model alone.
- Open Weights - AI models distributed with their trained parameters publicly available for download and use, without necessarily including the training data or full training code.
- AI Mixture of Experts - Architecture where multiple specialized sub-networks are selectively activated for different inputs to improve efficiency.
- Goldilocks Rule for AI - The principle that AI tasks should be neither too easy nor too hard to maintain engagement and optimal learning.
- Emergent Abilities - Capabilities that appear in large AI models only beyond a critical scale threshold, absent or near-random in smaller models.
- Autoregressive Model - A type of generative model that produces output sequentially, using each generated element as input for predicting the next one.
- Fine-Tuning - Customizing pre-trained AI models by training them further on specific data or tasks.
- Direct Preference Optimization - A simplified alternative to RLHF that fine-tunes language models directly on human preference data without training a separate reward model.
- AI Sampling Parameters - Configuration settings like temperature, top-p, and top-k that control the randomness and creativity of AI text generation.
- Reinforcement Learning from Human Feedback (RLHF) - A training technique that aligns LLM outputs with human preferences by using human feedback to guide model behavior.
- Automatic Speech Recognition - Technology that converts spoken language into text, enabling machines to understand and transcribe human speech.
- Edge AI - Running artificial intelligence models directly on local devices (phones, IoT sensors, cars) rather than in the cloud, enabling faster responses and greater privacy.
- Few-Shot Learning - Training or prompting AI with just a few examples to perform new tasks.
- AI Instruction Tuning - Training method that teaches AI models to follow natural language instructions by fine-tuning on instruction-response pairs.
- Algorithmic Bias - Systematic errors in AI and automated systems that create unfair outcomes, often reflecting or amplifying human biases present in training data or design choices.
- Instruction Tuning - A fine-tuning technique that trains language models to follow natural language instructions by learning from examples of instruction-response pairs.
- Perplexity - A measurement of how well a language model predicts text, with lower values indicating better performance and more confident predictions.
- Open Training - Practice of making the entire AI model training process transparent and reproducible, including training data, code, hyperparameters, and methodology.
- Speaker Diarization - The process of partitioning an audio stream into segments according to speaker identity, answering the question of 'who spoke when.'
- Dimensionality Reduction - A set of techniques for reducing the number of variables in a dataset while preserving its essential structure, making high-dimensional data easier to visualize, process, and analyze.
- Model Scaling - The study and practice of increasing neural network size, data, or compute to improve model performance, guided by empirical scaling laws.
- Reward Model - A neural network trained to predict human preferences, used to provide a scalar reward signal for optimizing language model behavior in RLHF.
- Model Pruning - A neural network compression technique that removes redundant or low-impact weights, neurons, or entire layers to create smaller, faster models.
- AI Foundation Models - Large-scale AI models trained on broad data that serve as the base for various downstream applications.
- AI KV Cache - Key-value caching mechanism that stores previously computed attention states to speed up sequential token generation.
- Reinforcement Learning - A machine learning paradigm where an agent learns to make decisions by taking actions in an environment and receiving rewards or penalties as feedback.
- AI Multimodal - AI systems that can process and generate multiple types of data including text, images, audio, and video.
- RAG Pipelines - Data processing workflows that handle the end-to-end flow from document ingestion to LLM response generation in Retrieval-Augmented Generation systems.
- AI Inference - The process of running a trained machine learning model to generate predictions, classifications, or outputs from new input data.
- Connectionism - Connectionism is a cognitive science approach that models mental processes using artificial neural networks of simple interconnected units processing information in parallel through weighted connections.
- Neural Scaling Laws - Empirical power-law relationships predicting how AI model performance improves as a function of model size, dataset size, and compute budget.
- Open-Source AI - Artificial intelligence systems released with open access to model weights, training code, data, and documentation, enabling community use, modification, and redistribution.
- AI Scaling Laws - Empirical relationships between model size, training data, compute, and AI performance that guide resource allocation.
- Reward Hacking - A failure mode in reinforcement learning where an agent exploits flaws in the reward function to achieve high reward without fulfilling the intended objective.
- Pre-training - The initial phase of training a language model on large-scale text data to learn general language understanding before task-specific fine-tuning.
- Machine Learning - A subset of artificial intelligence that enables systems to learn and improve from experience without being explicitly programmed.
- Vector Store - A specialized database designed to store, index, and search high-dimensional vector embeddings for AI applications.
- Next-Token Prediction - The core mechanism of autoregressive language models that generates text by predicting the most likely next token given all preceding tokens.
- AI Speculative Decoding - Technique where a smaller draft model generates candidate tokens that a larger model verifies in parallel to speed up inference.
- Mixture of Experts - A neural network architecture that uses a gating network to route inputs to specialized sub-networks called experts, enabling efficient scaling by activating only a subset of parameters for each input.
- Training Data - The dataset used to teach a machine learning model patterns and relationships, directly shaping the model's capabilities and limitations.
- AI Tokenization - Process of breaking text into tokens that AI models use as their fundamental units of input and output.
- Model Collapse - The degradation of AI model quality when trained on synthetic data generated by other AI models, causing progressive loss of diversity and accuracy.
- Knowledge Distillation - A model compression technique where a smaller student model is trained to reproduce the behavior and outputs of a larger, more capable teacher model.
- Unsupervised Learning - A machine learning approach where models find patterns in data without labeled examples or predefined outcomes.
- Zero-Shot Learning - AI performing tasks based on instructions alone, without any specific examples.
- AI Distillation - Training a smaller student model to replicate the behavior of a larger teacher model while maintaining performance.
- Representation Learning - A class of machine learning techniques where models automatically discover the representations needed for a task from raw data, rather than relying on manually engineered features.
- AI Open Weight Models - AI models whose trained parameters are publicly released, enabling local deployment, modification, and research.
- Deep Learning - A subset of machine learning using neural networks with multiple layers to learn complex patterns from data.
- Small Language Models (SLMs) - Compact language models optimized for efficiency that can run on consumer hardware while maintaining useful capabilities.
- Speculative Decoding - An inference acceleration technique where a smaller draft model proposes multiple tokens that a larger target model verifies in parallel, speeding up generation without changing output quality.
- Diffusion Models - Generative AI models that learn to create data by progressively denoising random noise into coherent outputs.
- Gating Network - A neural network component that learns to route inputs to the most appropriate expert sub-networks in mixture of experts architectures.
- AI Training Data Collection - The processes and ethical considerations of gathering data used to train AI models, including the use of user prompts and conversations as training signal.
- Federated Learning - A distributed machine learning approach where models are trained across multiple decentralized devices or servers holding local data, without exchanging raw data.
- Artificial Neural Network - A computing system inspired by biological neural networks that learns to perform tasks by processing examples through layers of interconnected nodes.
- AI Quantization - Reducing AI model precision from higher to lower bit representations to decrease size and increase speed.
- Natural Language Processing - The field of artificial intelligence focused on enabling computers to understand, interpret, and generate human language.
- Large Language Models (LLMs) - AI models that use transformer architecture to understand and generate human-like text by predicting the next token in a sequence.
- Latent Space - A compressed, multi-dimensional representation space where a model encodes the essential features of its input data.
- AI Fine-Tuning - Adapting a pre-trained AI model to a specific task or domain using additional targeted training.
- Model Quantization - A technique for reducing the numerical precision of a neural network's weights and activations to decrease model size, memory usage, and inference latency.
← Back to all concepts