UQLM provides a suite of response-level scorers for quantifying the uncertainty of Large Language Model (LLM) outputs. Each scorer returns a confidence score between 0 and 1, where higher scores ...
Abstract: Producing executable code from natural-language directives via Large Language Models (LLMs) involves obstacles like semantic uncertainty and the requirement for task-focused context ...
A general-purpose Claude Code action for GitHub PRs and issues that can answer questions and implement code changes. This action intelligently detects when to activate based on your workflow ...
Abstract: Recently, Large Language Models (LLMs) have made substantial progress in code generation, but they still frequently generate code containing logic errors or syntax bugs. While research has ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results