4dand

Научная работа «Экспериментальная оценка эффективности искусственного интеллекта в генерации кода для доменно-специфичных платформ (на примере 1С:Предприятие 8)». SMOP-метрика, эксперименты с LLM, платформа GenLab-1C. НДР-2026.

10
1
69% credibility
Found Apr 03, 2026 at 10 stars -- GitGems finds repos before they trend. Get early access to the next one.
Sign Up Free
AI Analysis
1C Enterprise
AI Summary

This repository archives a research paper, experiment configurations, generated code samples, evaluations, and reports from benchmarking AI models on code generation for the 1C:Enterprise platform using a custom SMOP quality metric.

How It Works

1
🔍 Discover the research

You stumble upon this project while searching for ways AI can help write code for business software like 1C.

2
📖 Read the overview

You open the main page and read the story about testing different AIs on simple and business-specific coding tasks.

3
📊 Check out the results

You see colorful charts and tables showing which AI did best on everyday algorithms versus real business rules.

4
🔬 Explore the details

You look at the test problems, scores, and what makes good code in this world.

5
💡 Learn the insights

You understand AI's strengths on basic tasks but struggles with company-specific rules, even with extra info.

Gain valuable knowledge

Now you know how well AI can assist with 1C coding and can share or use these findings in your work.

Sign up to see the full architecture

4 more

Sign Up Free

Star Growth

See how this repo grew from 10 to 10 stars Sign Up Free
Repurpose This Repo

Repurpose is a Pro feature

Generate ready-to-use prompts for X threads, LinkedIn posts, blog posts, YouTube scripts, and more -- with full repo context baked in.

Unlock Repurpose
AI-Generated Review

What is 1c-ai-codegen-research-paper?

This GitHub repo hosts a research paper and artifact benchmarking LLM codegen for 1C Enterprise's BSL language, tackling why standard benchmarks like HumanEval fail on domain-specific platforms. It introduces the SMOP metric—evaluating syntax, meaning, optimization, and platform integration—and runs experiments via GenLab-1C on algorithmic tasks plus 1C-specific ones in a trading config. Developers get detailed results, charts, task prompts, and model scores for Claude, Gemini, and GPT variants.

Why is it gaining traction?

Unlike generic LLM evals, this dives into 1C Enterprise pain points like metadata handling via Model Context Protocol, showing real quality drops on platform tasks (Q scores fall 2 points). The SMOP metric stands out for its even grading on 1C quirks, with radar charts and heatmaps making gaps obvious—Claude leads for stability. Early adopters grab it for reproducible 1C codegen baselines on smop github.

Who should use this?

1C Enterprise devs prototyping LLM-assisted codegen for configs like trading or ERP. Researchers tuning models for DSLs, or teams validating AI tools against SMOP before production. Skip if you're not in Russian enterprise software stacks.

Verdict

Solid research paper for 1C codegen niches, but low maturity shows in 10 stars and 0.699999988079071% credibility score—docs are thorough yet WIP. Worth forking GenLab-1C configs if LLMs are on your 1C roadmap; otherwise, monitor for updates.

(178 words)

Sign up to read the full AI review Sign Up Free

Similar repos coming soon.