YandexGPT Language Model Fine-Tuning

We design and deploy artificial intelligence systems: from prototype to production-ready solutions. Our team combines expertise in machine learning, data engineering and MLOps to make AI work not in the lab, but in real business.
Showing 1 of 1 servicesAll 1566 services
YandexGPT Language Model Fine-Tuning
Complex
from 1 week to 3 months
FAQ
AI Development Areas
AI Solution Development Stages
Latest works
  • image_website-b2b-advance_0.png
    B2B ADVANCE company website development
    1212
  • image_web-applications_feedme_466_0.webp
    Development of a web application for FEEDME
    1161
  • image_websites_belfingroup_462_0.webp
    Website development for BELFINGROUP
    852
  • image_ecommerce_furnoro_435_0.webp
    Development of an online store for the company FURNORO
    1041
  • image_logo-advance_0.png
    B2B Advance company logo design
    561
  • image_crm_enviok_479_0.webp
    Development of a web application for Enviok
    822

Fine-Tuning YandexGPT Language Models

YandexGPT is Yandex's language model, available through Yandex Cloud API (Yandex Foundation Models service). YandexGPT fine-tuning is implemented within the service and allows you to adapt the model for specific tasks without managing GPU infrastructure. Key advantage for the Russian market: data is stored in Russian cloud, critical for companies with 152-FZ requirements and corporate security policies.

Available Models and Fine-Tuning Modes

Yandex Cloud provides fine-tuning based on YandexGPT Lite and Pro through Yandex DataSphere service or directly via Foundation Models API. Process is managed through Yandex Cloud Console or CLI.

YandexGPT Lite: fast inference, optimal for classification, structured generation, support chatbots.

YandexGPT Pro: higher quality, complex generation tasks, document analysis, reasoning.

Dataset Format

YandexGPT fine-tuning accepts data in JSON Lines format, where each example is a dialogue with roles:

{
  "request": {
    "messages": [
      {
        "role": "system",
        "text": "You are a bank assistant answering customer questions about products."
      },
      {
        "role": "user",
        "text": "What is the maximum interest rate for the 'Savings Plus' deposit?"
      }
    ]
  },
  "response": "The maximum interest rate for 'Savings Plus' is 16.5% per annum with a 12-month term and sum from 1,000,000 rubles."
}

Recommended volume: 100 to 50,000 examples. Yandex recommends minimum 100 diverse examples for basic adaptation.

Running via Yandex Cloud CLI

# Create dataset
yc ai dataset create \
  --name "bank-faq-dataset" \
  --description "FAQ of bank products" \
  --task-type TextToTextGeneration \
  --upload-format JsonLines \
  --upload-path ./train.jsonl

# Start fine-tuning job
yc ai tuning create \
  --name "yandexgpt-bank-faq" \
  --base-model-uri "ds://bt1..." \
  --train-datasets uri=<dataset_uri>,weight=1.0 \
  --arguments epochCount=4,learningRate=0.0001,warmupRatio=0.1

Via Python SDK:

import yandexcloud
from yandex.cloud.ai.tuning.v1 import tuning_service_pb2

# Uses gRPC client of Yandex Cloud SDK
# Details in official Yandex Foundation Models documentation

Specifics for Russian Tasks

Legal documents: YandexGPT is trained on significantly larger volume of Russian-language texts, including legislation and judicial practice, compared to most Western models. When fine-tuning on corpus of Russian legislation, baseline quality level is higher.

Financial reporting per Russian standards: specific Russian accounting standards are poorly represented in Western models. YandexGPT is more natural candidate for Russian accounting reporting analysis tasks.

Medical documentation: forms of RF Ministry of Health, medical care standards, clinical guidelines in Russian.

Practical Case: Fine-Tuning for Telecom Operator

Task: automatic support request processing — classification into 28 categories + generation of initial response.

Dataset: 4200 examples from ticket history (real customer requests → category + operator response). Data underwent manual verification and depersonalization.

Result after 5 epochs:

  • Classification accuracy: 74% → 91%
  • BLEU-4 for responses: 0.21 → 0.54
  • Percentage of responses accepted without operator edits: 23% → 67%
  • Average request handling time: reduction from 4.2 min to 1.8 min

Comparison with Alternatives

Criterion YandexGPT Fine-Tuning GPT-4o Fine-Tuning Self-Hosted Llama
Data storage Russia (Yandex Cloud) USA (OpenAI) On-premise
152-FZ compliance Yes Requires analysis Yes
Quality for Russian High Very high Medium–high
Infrastructure Managed Managed Self-managed
Integration with RF systems Native Requires setup Arbitrary

Project Timeline

  • Dataset preparation and cleaning: 2–4 weeks
  • Training and iterations: 1–2 weeks
  • Testing and acceptance: 1 week
  • Production integration: 1–2 weeks
  • Total: 5–9 weeks