ai-engineer

Build production-ready LLM applications, advanced RAG systems, and intelligent agents. Implements vector search, multimodal AI, agent orchestration, and enterprise AI integrations. Use PROACTIVELY for LLM features, chatbots, AI agents, or AI-powered applications.

Author

Install

Hot:43

Download and extract to your skills directory

Copy command and send to OpenClaw for auto-install:

Download and install this skill https://openskills.cc/api/download?slug=sickn33-skills-ai-engineer&locale=en&source=copy

AI Engineer Skills

Skill Overview


Expert-level skills for building production-grade large language model applications, advanced RAG systems, and intelligent AI agents, covering vector search, multimodal AI, and enterprise AI integration.

Applicable Scenarios

  • Enterprise Knowledge Bases and Intelligent Q&A

  • Build RAG systems based on enterprise documents to achieve precise semantic retrieval and intelligent Q&A, supporting multiple document formats and hybrid search strategies.

  • AI Agents and Automated Workflows

  • Design multi-agent collaboration systems for complex task orchestration and tool invocation, suitable for customer service bots, research assistants, and business automation scenarios.

  • Production Deployment of LLM Applications

  • Integrate large language models into production environments, including cost optimization, performance tuning, security monitoring, and gradual rollout, ensuring system stability and reliability.

    Core Capabilities

  • Advanced RAG System Architecture

  • Support multi-stage retrieval pipelines, hybrid vector search, query optimization, and context compression. Integrate mainstream vector databases (Pinecone, Qdrant, Weaviate) to achieve high-accuracy knowledge retrieval.

  • AI Agent Framework Development

  • Build complex agent systems based on LangChain, LangGraph, LlamaIndex, and CrewAI, supporting multimodal tool invocation, memory management, and state orchestration.

  • Production-Grade AI System Operations

  • Provide complete monitoring, caching, degradation, and cost-control solutions, support streaming inference, A/B testing, and model routing to ensure stability for large-scale deployments.

    Frequently Asked Questions

    What scenarios are AI engineer skills suitable for?


    Suitable for any scenario that requires integration of large language models, including enterprise knowledge base Q&A, intelligent customer service, content generation, code assistants, and data analysis. Particularly strong for production-grade applications that need high-accuracy retrieval and complex reasoning.

    How do you build a production-grade RAG system?


    Core components include: document processing and chunking strategies, vector database selection, embedding model selection, hybrid retrieval (vector + keyword), re-ranking optimization, as well as query understanding and context compression. This skill provides a complete solution from architecture design to performance tuning.

    How can LLM applications control costs?


    Reduce costs through model selection strategies (combinations of GPT-4o/Claude/open-source models), semantic caching, prompt optimization, batching, and monitoring alerts. This skill includes cost-optimization best practices and supports automatically selecting the most economical model based on the scenario.

    Which vector database should I choose?


    Choose based on scale and requirements: Pinecone is suitable for quick deployments, Qdrant for self-hosting, Weaviate offers rich filtering capabilities, and pgvector fits scenarios with existing PostgreSQL. This skill supports all mainstream vector databases and can provide selection advice.

    How should multi-agent systems be designed?


    Adopt role division, collaborative orchestration, and shared memory patterns. Support a single agent handling simple tasks and multiple agents collaborating on complex workflows. Implement state management and task distribution based on LangGraph and CrewAI.

    How can AI applications ensure security?


    Built-in multi-layer security protections: content moderation, prompt injection detection, PII detection and redaction, access control, and audit logs. Support compliance checks and red-team testing to meet enterprise security requirements.

    How can RAG systems improve retrieval accuracy?


    Use hybrid retrieval (vector + BM25), query expansion and decomposition, re-ranking models, context compression, and negative feedback optimization. This skill supports advanced patterns like GraphRAG and HyDE to significantly improve retrieval quality.

    What are the best practices for deploying LLM applications in enterprises?


    Adopt progressive deployment: POC validation, small-scale pilots, gradual rollout, and full launch. Focus on observability, degradation strategies, cost monitoring, and model version management. This skill provides a complete production deployment framework.