Fix chat interface - restore continuous conversation flow

🎯 Major improvements to MissionControl component:
- Always keep input field visible and functional after AI responses
- Auto-clear input after submitting questions for better UX
- Add dynamic visual indicators (first question vs follow-up)
- Improve response layout with clear separation and hints
- Enable proper chat-like experience for continuous learning

🌟 Additional enhancements:
- Better language-specific messaging throughout interface
- Clearer visual hierarchy between input and response areas
- Intuitive flow that guides users to ask follow-up questions
- Maintains responsive design and accessibility

🔧 Technical changes:
- Enhanced MissionControl state management
- Improved component layout and styling
- Better TypeScript integration across components
- Updated tsconfig for stricter type checking
This commit is contained in:
rwiegand
2025-07-14 12:39:05 +02:00
parent b31492a354
commit f893530471
1798 changed files with 25329 additions and 92638 deletions

View File

@@ -1,7 +1,6 @@
import { APIResource } from "../../../core/resource.js";
import { APIResource } from "../../../resource.js";
import * as Core from "../../../core.js";
import * as GraderModelsAPI from "../../graders/grader-models.js";
import { APIPromise } from "../../../core/api-promise.js";
import { RequestOptions } from "../../../internal/request-options.js";
export declare class Graders extends APIResource {
/**
* Run a grader.
@@ -17,10 +16,11 @@ export declare class Graders extends APIResource {
* type: 'string_check',
* },
* model_sample: 'model_sample',
* reference_answer: 'string',
* });
* ```
*/
run(body: GraderRunParams, options?: RequestOptions): APIPromise<GraderRunResponse>;
run(body: GraderRunParams, options?: Core.RequestOptions): Core.APIPromise<GraderRunResponse>;
/**
* Validate a grader.
*
@@ -38,17 +38,13 @@ export declare class Graders extends APIResource {
* });
* ```
*/
validate(body: GraderValidateParams, options?: RequestOptions): APIPromise<GraderValidateResponse>;
validate(body: GraderValidateParams, options?: Core.RequestOptions): Core.APIPromise<GraderValidateResponse>;
}
export interface GraderRunResponse {
metadata: GraderRunResponse.Metadata;
model_grader_token_usage_per_model: {
[key: string]: unknown;
};
model_grader_token_usage_per_model: Record<string, unknown>;
reward: number;
sub_rewards: {
[key: string]: unknown;
};
sub_rewards: Record<string, unknown>;
}
export declare namespace GraderRunResponse {
interface Metadata {
@@ -56,9 +52,7 @@ export declare namespace GraderRunResponse {
execution_time: number;
name: string;
sampled_model_name: string | null;
scores: {
[key: string]: unknown;
};
scores: Record<string, unknown>;
token_usage: number | null;
type: string;
}
@@ -93,19 +87,13 @@ export interface GraderRunParams {
*/
grader: GraderModelsAPI.StringCheckGrader | GraderModelsAPI.TextSimilarityGrader | GraderModelsAPI.PythonGrader | GraderModelsAPI.ScoreModelGrader | GraderModelsAPI.MultiGrader;
/**
* The model sample to be evaluated. This value will be used to populate the
* `sample` namespace. See
* [the guide](https://platform.openai.com/docs/guides/graders) for more details.
* The `output_json` variable will be populated if the model sample is a valid JSON
* string.
* The model sample to be evaluated.
*/
model_sample: string;
/**
* The dataset item provided to the grader. This will be used to populate the
* `item` namespace. See
* [the guide](https://platform.openai.com/docs/guides/graders) for more details.
* The reference answer for the evaluation.
*/
item?: unknown;
reference_answer: string | unknown | Array<unknown> | number;
}
export interface GraderValidateParams {
/**