温馨提示:本站仅提供公开网络链接索引服务,不存储、不篡改任何第三方内容,所有内容版权归原作者所有
AI智能索引来源:http://www.ibm.com/think/topics/llm-alignment
点击访问原文链接

What Is LLM Alignment? | IBM

HomeThinkTopicsWelcomeOverviewMachine learning typesMachine learning algorithmsStatistical machine learningLinear algebra for machine learningUncertainty quantificationBias variance tradeoffBayesian StatisticsSingular value decompositionOverviewFeature selectionFeature extractionVector embeddingLatent spacePrincipal component analysisLinear discriminant analysisUpsamplingDownsamplingSynthetic dataData leakageOverviewLinear regressionLasso regressionRidge regressionState space modelTime seriesAutoregressive modelOverviewDecision treesK-nearest neighbors (KNNs)Naive bayesRandom forestSupport vector machineLogistic regressionOverviewBoostingBaggingGradient boostingGradient boosting classifierOverviewTransfer learningOverviewOverviewK means clusteringHierarchical clusteringA priori algorithmGaussian mixture modelAnomaly detectionOverviewCollaborative filteringContent based filteringOverviewReinforcement learning human feedbackOverviewOverviewBackpropagationEncoder-decoder modelRecurrent neural networksLong short-term memory (LSTM)Convolutional neural networksOverviewAttention mechanismGrouped query attentionPositional encodingAutoencoderMamba modelGraph neural networkOverviewGenerative modelGenerative AI vs. predictive AIOverviewReasoning modelsSmall language modelsInstruction tuningLLM parametersLLM temperatureLLM benchmarksLLM customizationLLM alignmentDiffusion modelsVariational autoencoder (VAE)Generative adversarial networks (GANs)OverviewVision language modelsTutorial: Build an AI stylistTutorial: Multimodal AI queries using LlamaTutorial: Multimodal AI queries using PixtralTutorial: Automatic podcast transcription with GraniteTutorial: PPT AI image analysis answering systemOverviewGraphRAGTutorial: Build a multimodal RAG system with Docling and GraniteTutorial: Evaluate RAG pipline using RagasTutorial: RAG chunking strategiesTutorial: Graph RAG using knowledge graphsTutorial: Inference scaling to improve multimodal RAGOverviewVibe codingVisit the 2025 Guide to AI AgentsLLM trainingOverviewLoss functionTraining dataModel parametersGradient descentStochastic gradient descentHyperparameter tuningLearning rateOverviewParameter efficient fine tuning (PEFT)LoRATutorial: Fine tuning Granite model with LoRARegularizationFoundation modelsOverfittingUnderfittingFew shot learningZero shot learningKnowledge distillationMeta learningData augmentationCatastrophic forgettingOverviewScikit-learnXGboostPyTorchOverviewAI lifecyleAI inferenceModel deploymentMachine learning pipelineData labelingModel risk managementModel driftAutoMLModel selectionFederated learningDistributed machine learningAI stackOverviewNatural language understandingOverviewSentiment analysisTutorial: Spam text classifier with PyTorchMachine translationOverviewInformation retrievalInformation extractionTopic modelingLatent semantic analysisLatent Dirichlet AllocationNamed entity recognitionWord embeddingsBag of wordsIntelligent searchSpeech recognitionStemming and lemmatizationText summarizationConversational AIConversational analyticsNatural language generationOverviewImage classificationObject detectionInstance segmentationSemantic segmentationOptical character recognitionImage recognitionVisual inspectionDave Bergmannlarge language model (LLM)alignedpretrainingfine-tuningagentic AIartificial intelligenceAI safetyresponsiblyartificial general intelligence (AGI)riskstrainself-supervisedtraining datadatasetscrapingAI hallucinationsguardrailsprovoke nuclear warIBM Privacy Statementtraining processneural networkfine-tuninginstruction tuningSystem promptssupervised learningreinforcement learningGrade School Math 8K (GSM8K)trained entirely on enterprise-safe dataneural network“abliteration.”experimental LLM architecture from Guide Labsreasoning modelsreasoning models aren’t always “honest” when verbalizing their chain of thoughtGo to episodeAnthropic’s Claude AIAPIchatbotprompt injection attacksinstruction tuningcontext lengthmodel parametersloss functionknowledge distillationsynthetic datajailbreakingreinforcement learning from human feedback (RLHF)algorithmsReinforcement learning from human feedback (RLHF)ChatGPTproximal policy optimization (PPO).It can also lead to sycophancyconstitutional AI“Safety Pretraining: Toward the Next Generation of Safe AI,”classifierinferencebeam searchTruthfulQAHarmBenchChatbotArenaRed teamingDave BergmannUpcoming Webinar | March 31 Achieve continuous compliance in a hybrid data world with IBM Guardium Data Protection Register for this webinar to learn how AI governance helps organizations manage risk, meet evolving regulations and build trusted, responsible AI at scale. Register nowReport IBM X-Force Threat Intelligence Index 2026 Gain insights to prepare and respond to cyberattacks with greater speed and effectiveness with the IBM X-Force® Threat Intelligence Index. Read the reportReport AI governance imperative: evolving regulations and emergence of agentic AI Learn how evolving regulations and the emergence of AI agents are reshaping the need for robust AI governance frameworks. Read the reportWebinar recording Agent Ops and Responsible AI Join this webinar to explore practical strategies for operating and governing AI agents responsibly at scale, with expert insights on observability, risk management and accountable AI operations. Watch nowGuide Building a strong data foundation for trustworthy AI Explore the Data Matters hub to see how strong data practices and governance lay the foundation for scalable AI success. See why Data MattersEbook Maximize AI ROI through smarter governance Learn ways to maximize AI ROI—prioritizing high-impact use cases, governing risks, optimizing costs and accelerating adoption with watsonx®. Read the ebookReport IBM named a Leader in the Gartner® Magic Quadrant™ for GRC Unlock insights into IBM's OpenPages and learn why we were named a Leader. Get the ReportReport Why AI governance is a business imperative for scaling enterprise artificial intelligence Learn about the new challenges of generative AI, the need for governing AI and ML models and steps to build a trusted, transparent and explainable AI framework. Read the reportReport Getting ready for the EU AI Act, Phase 2: Risk-Assess and Categorize Understand the importance of establishing a defensible assessment process and consistently categorizing each use case into the appropriate risk tier. Read the reportInsight AI lifecycle governance Read about driving ethical and compliant practices with a portfolio of AI products for generative AI models. Read the insightEbook How to choose the right foundation model Learn how to select the most suitable AI foundation model for your use case. Read the ebookDiscover watsonx.governanceDiscover AI governance solutionsDiscover AI governance servicesExplore watsonx.governanceBook a live demo“A General Language Assistant as a Laboratory for Alignment,”“Ethical Issues in Advanced Artificial Intelligence,”“Safety Pretraining: Toward the Next Generation of Safe AI,”“Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs,”“Safety Alignment Should Be made More Than Just a Few Tokens Deep,”“Refusal in LLMs is mediated by a single direction,”“Unpacking Claude’s System Prompt,”“Is DPO Superior to PPO for LLM Alignment? A Comprehensive Study,”

智能索引记录