Files
kidsai/node_modules/openai/resources/fine-tuning/methods.d.ts
rwiegand f893530471 Fix chat interface - restore continuous conversation flow
🎯 Major improvements to MissionControl component:
- Always keep input field visible and functional after AI responses
- Auto-clear input after submitting questions for better UX
- Add dynamic visual indicators (first question vs follow-up)
- Improve response layout with clear separation and hints
- Enable proper chat-like experience for continuous learning

🌟 Additional enhancements:
- Better language-specific messaging throughout interface
- Clearer visual hierarchy between input and response areas
- Intuitive flow that guides users to ask follow-up questions
- Maintains responsive design and accessibility

🔧 Technical changes:
- Enhanced MissionControl state management
- Improved component layout and styling
- Better TypeScript integration across components
- Updated tsconfig for stricter type checking
2025-07-14 12:39:05 +02:00

120 lines
4.0 KiB
TypeScript

import { APIResource } from "../../resource.js";
import * as GraderModelsAPI from "../graders/grader-models.js";
export declare class Methods extends APIResource {
}
/**
* The hyperparameters used for the DPO fine-tuning job.
*/
export interface DpoHyperparameters {
/**
* Number of examples in each batch. A larger batch size means that model
* parameters are updated less frequently, but with lower variance.
*/
batch_size?: 'auto' | number;
/**
* The beta value for the DPO method. A higher beta value will increase the weight
* of the penalty between the policy and reference model.
*/
beta?: 'auto' | number;
/**
* Scaling factor for the learning rate. A smaller learning rate may be useful to
* avoid overfitting.
*/
learning_rate_multiplier?: 'auto' | number;
/**
* The number of epochs to train the model for. An epoch refers to one full cycle
* through the training dataset.
*/
n_epochs?: 'auto' | number;
}
/**
* Configuration for the DPO fine-tuning method.
*/
export interface DpoMethod {
/**
* The hyperparameters used for the DPO fine-tuning job.
*/
hyperparameters?: DpoHyperparameters;
}
/**
* The hyperparameters used for the reinforcement fine-tuning job.
*/
export interface ReinforcementHyperparameters {
/**
* Number of examples in each batch. A larger batch size means that model
* parameters are updated less frequently, but with lower variance.
*/
batch_size?: 'auto' | number;
/**
* Multiplier on amount of compute used for exploring search space during training.
*/
compute_multiplier?: 'auto' | number;
/**
* The number of training steps between evaluation runs.
*/
eval_interval?: 'auto' | number;
/**
* Number of evaluation samples to generate per training step.
*/
eval_samples?: 'auto' | number;
/**
* Scaling factor for the learning rate. A smaller learning rate may be useful to
* avoid overfitting.
*/
learning_rate_multiplier?: 'auto' | number;
/**
* The number of epochs to train the model for. An epoch refers to one full cycle
* through the training dataset.
*/
n_epochs?: 'auto' | number;
/**
* Level of reasoning effort.
*/
reasoning_effort?: 'default' | 'low' | 'medium' | 'high';
}
/**
* Configuration for the reinforcement fine-tuning method.
*/
export interface ReinforcementMethod {
/**
* The grader used for the fine-tuning job.
*/
grader: GraderModelsAPI.StringCheckGrader | GraderModelsAPI.TextSimilarityGrader | GraderModelsAPI.PythonGrader | GraderModelsAPI.ScoreModelGrader | GraderModelsAPI.MultiGrader;
/**
* The hyperparameters used for the reinforcement fine-tuning job.
*/
hyperparameters?: ReinforcementHyperparameters;
}
/**
* The hyperparameters used for the fine-tuning job.
*/
export interface SupervisedHyperparameters {
/**
* Number of examples in each batch. A larger batch size means that model
* parameters are updated less frequently, but with lower variance.
*/
batch_size?: 'auto' | number;
/**
* Scaling factor for the learning rate. A smaller learning rate may be useful to
* avoid overfitting.
*/
learning_rate_multiplier?: 'auto' | number;
/**
* The number of epochs to train the model for. An epoch refers to one full cycle
* through the training dataset.
*/
n_epochs?: 'auto' | number;
}
/**
* Configuration for the supervised fine-tuning method.
*/
export interface SupervisedMethod {
/**
* The hyperparameters used for the fine-tuning job.
*/
hyperparameters?: SupervisedHyperparameters;
}
export declare namespace Methods {
export { type DpoHyperparameters as DpoHyperparameters, type DpoMethod as DpoMethod, type ReinforcementHyperparameters as ReinforcementHyperparameters, type ReinforcementMethod as ReinforcementMethod, type SupervisedHyperparameters as SupervisedHyperparameters, type SupervisedMethod as SupervisedMethod, };
}
//# sourceMappingURL=methods.d.ts.map