• Home
  • Categories
  • Pricing
  • Submit
    Built with
    Ever Works
    Ever Works

    Connect with us

    Stay Updated

    Get the latest updates and exclusive content delivered to your inbox.

    Product

    • Categories
    • Pricing
    • Help

    Clients

    • Sign In
    • Register
    • Forgot password?

    Company

    • About Us
    • Admin
    • Sitemap

    Resources

    • Blog
    • Submit
    • API Documentation
    All product names, logos, and brands are the property of their respective owners. All company, product, and service names used in this repository, related repositories, and associated websites are for identification purposes only. The use of these names, logos, and brands does not imply endorsement, affiliation, or sponsorship. This directory may include content generated by artificial intelligence.
    Copyright © 2025 Ever. All rights reserved.·Terms of Service·Privacy Policy·Cookies
    Decorative pattern
    Decorative pattern
    1. Home
    2. Machine Learning & Ai
    3. Awesome AI Engineering

    Awesome AI Engineering

    The Full-Stack LLM Engineering Playbook featuring architectural patterns for AI Agents with MCP and RAG, coupled with advanced post-training recipes including SFT, DPO, and QLoRA for domain adaptation, covering data pipelines, evaluation frameworks, and system design.

    Overview

    Awesome AI Engineering provides a comprehensive playbook for building production-grade LLM applications and AI systems. This resource covers the full stack of AI engineering, from model fine-tuning and post-training to system architecture, evaluation, and deployment at scale.

    Features

    • Architectural Patterns: Design patterns for AI agents and RAG systems
    • MCP Integration: Model Context Protocol for agent development
    • Post-Training Techniques: SFT, DPO, QLoRA, and other fine-tuning methods
    • Data Pipelines: Building and managing training data workflows
    • Evaluation Frameworks: Comprehensive model and system evaluation
    • System Design: Production architecture for LLM applications
    • Domain Adaptation: Techniques for specializing models
    • Deployment Strategies: Serving and scaling LLM systems

    AI Agent Architecture

    Agent Patterns

    • ReAct: Reasoning and Acting framework
    • Chain-of-Thought: Step-by-step reasoning
    • Tree of Thoughts: Exploring multiple reasoning paths
    • Reflection: Self-critique and improvement
    • Multi-Agent Systems: Collaborative agent architectures

    MCP (Model Context Protocol)

    • Standardized context management
    • Tool integration patterns
    • Memory management
    • State persistence
    • Agent communication protocols

    Tool Integration

    • Function calling patterns
    • Tool selection strategies
    • Error handling and retries
    • Rate limiting and quotas
    • Tool result parsing

    RAG (Retrieval-Augmented Generation)

    Architecture Patterns

    • Basic RAG: Simple retrieval and generation
    • Hybrid Search: Combining dense and sparse retrieval
    • Multi-Query RAG: Query expansion and reformulation
    • Parent-Child Chunking: Hierarchical document structure
    • Re-ranking: Improving retrieval relevance

    Retrieval Strategies

    • Vector search with embeddings
    • BM25 keyword search
    • Hybrid fusion methods
    • Contextual compression
    • Query routing

    Optimization Techniques

    • Chunk size optimization
    • Embedding model selection
    • Context window management
    • Cache strategies
    • Latency optimization

    Post-Training Recipes

    Supervised Fine-Tuning (SFT)

    • Dataset preparation and curation
    • Training hyperparameters
    • Learning rate schedules
    • Batch size optimization
    • Overfitting prevention

    Direct Preference Optimization (DPO)

    • Preference data collection
    • Reward model training
    • Policy optimization
    • Alignment techniques
    • Safety considerations

    QLoRA (Quantized LoRA)

    • 4-bit quantization
    • LoRA adapter training
    • Memory-efficient fine-tuning
    • Merge and deployment
    • Performance vs. efficiency trade-offs

    Domain Adaptation

    • Continued pre-training
    • Task-specific fine-tuning
    • Few-shot adaptation
    • Transfer learning strategies
    • Multi-task learning

    Data Pipelines

    Data Collection

    • Web scraping strategies
    • Synthetic data generation
    • Data augmentation
    • Active learning
    • Human-in-the-loop labeling

    Data Processing

    • Cleaning and deduplication
    • Quality filtering
    • Format standardization
    • Tokenization strategies
    • Dataset versioning

    Data Quality

    • Quality metrics
    • Outlier detection
    • Bias detection
    • Diversity measurement
    • Validation protocols

    Evaluation Frameworks

    Model Evaluation

    • Perplexity and loss metrics
    • Task-specific benchmarks
    • Human evaluation
    • A/B testing
    • Safety evaluations

    System Evaluation

    • End-to-end latency
    • Throughput measurement
    • Cost per request
    • Success rate tracking
    • User satisfaction metrics

    LLM-as-Judge

    • Automated evaluation with LLMs
    • Rubric-based assessment
    • Pairwise comparison
    • Multi-dimensional scoring

    System Design

    Architecture Patterns

    • Microservices for LLM apps
    • Event-driven architectures
    • Async processing
    • Queue management
    • State management

    Scalability

    • Horizontal scaling strategies
    • Load balancing
    • Caching layers
    • Database optimization
    • CDN integration

    Reliability

    • Error handling and retries
    • Circuit breakers
    • Fallback strategies
    • Monitoring and alerting
    • Disaster recovery

    Deployment and Serving

    Model Serving

    • vLLM for high-throughput
    • TGI (Text Generation Inference)
    • Ray Serve for distributed
    • Triton Inference Server
    • Custom serving solutions

    Optimization

    • Quantization (INT8, INT4)
    • Model pruning
    • Knowledge distillation
    • Speculative decoding
    • Continuous batching

    Infrastructure

    • GPU selection and optimization
    • Kubernetes deployment
    • Serverless options
    • Multi-region deployment
    • Cost optimization

    Best Practices

    Development

    • Start with prompting before fine-tuning
    • Use smaller models when possible
    • Implement comprehensive logging
    • Version everything (models, data, code)
    • Build evaluation early

    Production

    • Monitor model performance continuously
    • Implement rate limiting
    • Handle failures gracefully
    • Collect user feedback
    • Plan for model updates

    Ethics and Safety

    • Content filtering
    • Bias mitigation
    • Privacy protection
    • Toxicity detection
    • User consent management

    Pricing

    Free and open-source resource.

    Surveys

    Loading more......

    Information

    Websitegithub.com
    PublishedMar 22, 2026

    Categories

    1 Item
    Machine Learning & Ai

    Tags

    3 Items
    #ai-engineering#llm#mlops

    Similar Products

    6 result(s)

    Awesome LLMOps

    A curated collection of tools, frameworks, platforms, and best practices for operationalizing Large Language Models, covering deployment, monitoring, evaluation, and production workflows.

    Featured

    Awesome LangChain

    A curated collection of tools, projects, tutorials, and resources for LangChain, the popular framework for developing applications powered by large language models through composable components.

    Awesome Production Machine Learning

    A curated list of awesome open source libraries to deploy, monitor, version, and scale machine learning systems in production. Focuses on MLOps, model serving, and production-ready ML tools.

    Awesome Feature Store

    A curated list of open source and commercial feature store tools and platforms for machine learning, including Feast, Hopsworks, and Tecton, enabling consistent feature management across training and serving for production ML systems.

    Awesome LLM Prompt Optimization

    A curated list of advanced prompt optimization and tuning methods in Large Language Models.

    Awesome LLM Resources

    A comprehensive collection of the world's best LLM resources covering multimodal generation, AI agents, programming assistance, AI paper review, data processing, model training, model inference, o1 models, MCP, small language models, and vision-language models.