温馨提示:本站仅提供公开网络链接索引服务,不存储、不篡改任何第三方内容,所有内容版权归原作者所有
AI智能索引来源:http://www.ibm.com/think/topics/reasoning-model
点击访问原文链接

What Is a Reasoning Model? | IBM

WelcomeOverviewMachine learning typesMachine learning algorithmsStatistical machine learningLinear algebra for machine learningData visualization for machine learningUncertainty quantificationBias variance tradeoffBayesian StatisticsSingular value decompositionOverviewFeature selectionFeature extractionVector embeddingLatent spacePrincipal component analysisLinear discriminant analysisUpsamplingDownsamplingSynthetic dataData leakageOverviewLinear regressionLasso regressionRidge regressionState space modelTime seriesAutoregressive modelOverviewDecision treesK-nearest neighbors (KNNs)Naive bayesRandom forestSupport vector machineLogistic regressionOverviewBoostingBaggingGradient boostingGradient boosting classifierOverviewTransfer learningOverviewOverviewK means clusteringHierarchical clusteringA priori algorithmGaussian mixture modelAnomaly detectionOverviewCollaborative filteringContent based filteringOverviewReinforcement learning human feedbackDeep reinforcement learningOverviewOverviewBackpropagationEncoder-decoder modelRecurrent neural networksLong short-term memory (LSTM)Convolutional neural networksOverviewAttention mechanismGrouped query attentionPositional encodingAutoencoderMamba modelGraph neural networkOverviewGenerative modelGenerative AI vs. predictive AIOverviewReasoning modelsSmall language modelsInstruction tuningLLM parametersLLM temperatureLLM benchmarksLLM customizationLLM alignmentTutorial: Multilingual LLM agentDiffusion modelsVariational autoencoder (VAE)Generative adversarial networks (GANs)OverviewVision language modelsTutorial: Build an AI stylistTutorial: Multimodal AI queries using LlamaTutorial: Multimodal AI queries using PixtralTutorial: Automatic podcast transcription with GraniteTutorial: PPT AI image analysis answering systemOverviewGraphRAGTutorial: Build a multimodal RAG system with Docling and GraniteTutorial: Evaluate RAG pipline using RagasTutorial: RAG chunking strategiesTutorial: Graph RAG using knowledge graphsTutorial: Inference scaling to improve multimodal RAGOverviewVibe codingVisit the 2025 Guide to AI AgentsOverviewLLM trainingLoss functionTraining dataModel parametersOverviewGradient descentStochastic gradient descentHyperparameter tuningLearning rateOverviewParameter efficient fine tuning (PEFT)LoRATutorial: Fine tuning Granite model with LoRARegularizationFoundation modelsOverfittingUnderfittingFew shot learningZero shot learningKnowledge distillationMeta learningData augmentationCatastrophic forgettingOverviewScikit-learnXGboostPyTorchOverviewAI lifecyleAI inferenceModel deploymentMachine learning pipelineData labelingModel risk managementModel driftAutoMLModel selectionFederated learningDistributed machine learningAI stackOverviewNatural language understandingOverviewSentiment analysisTutorial: Spam text classifier with PyTorchMachine translationOverviewInformation retrievalInformation extractionTopic modelingLatent semantic analysisLatent Dirichlet AllocationNamed entity recognitionWord embeddingsBag of wordsIntelligent searchSpeech recognitionStemming and lemmatizationText summarizationConversational AIConversational analyticsNatural language generationOverviewImage classificationObject detectionInstance segmentationSemantic segmentationOptical character recognitionImage recognitionVisual inspectionDave Bergmannlarge language model (LLM)fine-tunedtraining modelsmachine learningtraining dataartificial general intelligence (AGI)DeepSeek-R1IBM GraniteIBM Privacy Statement“System 1” and “System 2” thinkingchain of thought promptinginference scalingself-supervisedsupervisedinstruction-tunedReinforcement learning from human feedback (RLHF)loss functionDeepSeek-R1 modelbackpropagatedsynthetic training dataKnowledge distillationfew-shotGo to episodeArenaHardAlpaca-Eval-2thought preference optimization (TPO)context windowAPIUpcoming webinar - May 19, 2026 Designing customer engagement: AI workflows that solve problems, not frustrate customers Learn how AI agents reduce service friction, fit into real workflows, and help teas deliver faster, more engaging customer experiences. Join nowGuide The CEO's guide to model optimization Learn how to continually push teams to improve model performance and outpace the competition by using the latest AI techniques and infrastructure. Read the guideTraining watsonx® Developer Hub Support your next project with some of our most commonly used capabilities. Get started and learn more about the supported models that IBM provides. Get startedReport A differentiated approach to AI foundation models Explore the value of enterprise-grade foundation models that provide trust, performance and cost-effective benefits to all industries. Read the reportEbook Unlock the power of generative AI and ML Learn how to incorporate generative AI, machine learning and foundation models into your business operations for improved performance. Read the ebookInsight How IBM is tailoring generative AI for enterprises Learn how IBM is developing generative foundation models that are trustworthy, energy efficient and portable. Read the insightExplore GraniteExplore AI solutionsExplore AI servicesDiscover watsonx.aiExplore IBM Granite AI models"The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity,""Introducing OpenAI o1-preview,""From System 1 to System 2: A Survey of Reasoning Large Language Models,""Large Language Models are Zero-Shot Reasoners,""Show Your Work: Scratchpads for Intermediate Computation with Language Models,""Scaling LLM Test-Time Compute Optimally can be More Effective than Scaling Model Parameters,""Let's Verify Step by Step,""Math-Shepherd: Verify and Reinforce LLMs Step-by-step without Human Annotations,""s1: Simple test-time scaling,""Inference Scaling Laws: An Empirical Analysis of Compute-Optimal Inference for Problem-Solving with Language Models,""STaR: Bootstrapping Reasoning With Reasoning,""Reinforced Self-Training (ReST) for Language Modeling,""Do NOT Think That Much for 2+3=? On the Overthinking of o1-Like LLMs,""The Danger of Overthinking: Examining the Reasoning-Action Dilemma in Agentic Tasks,""Inverse Scaling in Test-Time Compute,""Bringing reasoning to Granite,""Claude 3.7 Sonnet and Claude Code,""Generative AI on Vertex AI: Thinking,""Reasoning models don't always say what they think,"

智能索引记录