Skip to content

Commit 7cb333f

Browse files
authored
Core/clean codebase (#225)
1 parent a78bdb5 commit 7cb333f

File tree

11 files changed

+28
-28
lines changed

11 files changed

+28
-28
lines changed

packages/agent/src/agents/operators/supervisor/tools/schemas/common.schemas.ts

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ const modelGuardsValues: GuardsConfig['agents']['graph']['model'] =
3636
graphGuardsValues.model;
3737
// Schema for ModelConfig
3838
export const ModelConfigSchema = z.object({
39-
provider: z
39+
model_provider: z
4040
.string()
4141
.refine(
4242
(val) =>

packages/agent/src/agents/operators/supervisor/utils/normalizeAgentValues.ts

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -110,19 +110,19 @@ function normalizeModelConfig(
110110

111111
if (model && isPlainObject(model)) {
112112
const config: AgentConfig.Input['graph']['model'] = {
113-
provider: DEFAULT_AGENT_CONFIG.graph.model.provider,
113+
model_provider: DEFAULT_AGENT_CONFIG.graph.model.model_provider,
114114
model_name: DEFAULT_AGENT_CONFIG.graph.model.model_name,
115115
temperature: DEFAULT_AGENT_CONFIG.graph.model.temperature,
116116
max_tokens: DEFAULT_AGENT_CONFIG.graph.model.max_tokens,
117117
};
118118

119119
// Normalize provider
120120
const providerResult = normalizeStringValue(
121-
model.provider,
122-
DEFAULT_AGENT_CONFIG.graph.model.provider,
123-
'model.provider'
121+
model.model_provider,
122+
DEFAULT_AGENT_CONFIG.graph.model.model_provider,
123+
'model.model_provider'
124124
);
125-
config.provider = providerResult.value;
125+
config.model_provider = providerResult.value;
126126
if (providerResult.appliedDefault) {
127127
appliedDefaults.push(providerResult.appliedDefault);
128128
}

packages/core/src/common/agent/interfaces/agent.interface.ts

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ import { BaseChatModel } from '@langchain/core/language_models/chat_models';
33
export type Modify<T, R> = Omit<T, keyof R> & R;
44

55
export interface ModelConfig {
6-
provider: string;
6+
model_provider: string;
77
model_name: string;
88
description?: string;
99
temperature: number;

packages/core/src/common/constant/agents.constants.ts

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ export const agentSelectorConfig: AgentConfig.Input = {
2828
execution_timeout_ms: 30000,
2929
max_token_usage: 4000,
3030
model: {
31-
provider: 'gemini',
31+
model_provider: 'gemini',
3232
model_name: 'gemini-2.5-flash',
3333
temperature: 0.7,
3434
max_tokens: 8192,
@@ -90,7 +90,7 @@ export const supervisorAgentConfig: AgentConfig.Input = {
9090
execution_timeout_ms: 120000,
9191
max_token_usage: 16000,
9292
model: {
93-
provider: 'gemini',
93+
model_provider: 'gemini',
9494
model_name: 'gemini-2.5-flash',
9595
temperature: 0.7,
9696
max_tokens: 8192,
@@ -135,7 +135,7 @@ export namespace AgentConfigDefaults {
135135
execution_timeout_ms: 120000,
136136
max_token_usage: 16000,
137137
model: {
138-
provider: 'gemini',
138+
model_provider: 'gemini',
139139
model_name: 'gemini-2.5-flash',
140140
temperature: 0.7,
141141
max_tokens: 8192,

packages/core/src/common/constant/default-agent.constant.ts

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ import {
55
} from '@common/agent/interfaces/agent.interface.js';
66

77
export const DEFAULT_AGENT_MODEL: ModelConfig = {
8-
provider: 'gemini',
8+
model_provider: 'gemini',
99
model_name: 'gemini-2.5-flash',
1010
temperature: 0.7,
1111
max_tokens: 4096,

packages/core/src/services/agent-validation.service.ts

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -275,10 +275,10 @@ export class AgentValidationService {
275275
const maxTokens = getGuardValue('agents.graph.model.max_tokens');
276276

277277
// Validate provider - REQUIRED
278-
if (!model.provider) {
278+
if (!model.model_provider) {
279279
throw new Error('Model provider is required');
280280
}
281-
if (model.provider.toLowerCase() !== allowedProvider.toLowerCase()) {
281+
if (model.model_provider.toLowerCase() !== allowedProvider.toLowerCase()) {
282282
throw new Error(
283283
`Invalid model provider. Only '${allowedProvider}' is supported.`
284284
);

packages/database/__test__/database.spec.ts

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -532,21 +532,22 @@ describe('Custom Types Operations', () => {
532532

533533
it('should work with model composite type', async () => {
534534
const q = new Postgres.Query(
535-
`SELECT ROW('openai', 'gpt-4', 'Advanced AI model')::model as model_config,
536-
(ROW('openai', 'gpt-4', 'Advanced AI model')::model).provider as provider_name`
535+
`SELECT ROW('gemini', 'gemini-2.5-flash', 0.7, 4096)::model_config as model_config,
536+
(ROW('gemini', 'gemini-2.5-flash', 0.7, 4096)::model_config).model_provider as provider_name`
537537
);
538538

539539
interface ModelTest {
540540
model_config: {
541-
provider: string;
542-
modelName: string;
543-
description: string;
541+
model_provider: string;
542+
model_name: string;
543+
temperature: number;
544+
max_tokens: number;
544545
};
545546
provider_name: string;
546547
}
547548

548549
const [result] = await Postgres.query<ModelTest>(q);
549-
expect(result.provider_name).toBe('openai');
550+
expect(result.provider_name).toBe('gemini');
550551
});
551552
});
552553

packages/database/initdb/03-agents.sql

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -478,7 +478,7 @@ BEGIN
478478
COALESCE((p_config->'graph'->>'execution_timeout_ms')::bigint, (graph).execution_timeout_ms),
479479
COALESCE((p_config->'graph'->>'max_token_usage')::integer, (graph).max_token_usage),
480480
ROW(
481-
COALESCE(p_config->'graph'->'model'->>'provider', ((graph).model).model_provider),
481+
COALESCE(p_config->'graph'->'model'->>'model_provider', p_config->'graph'->'model'->>'provider', ((graph).model).model_provider),
482482
COALESCE(p_config->'graph'->'model'->>'model_name', ((graph).model).model_name),
483483
COALESCE((p_config->'graph'->'model'->>'temperature')::numeric(3,2), ((graph).model).temperature),
484484
COALESCE((p_config->'graph'->'model'->>'max_tokens')::integer, ((graph).model).max_tokens)
@@ -700,7 +700,7 @@ BEGIN
700700
(p_config->'graph'->>'execution_timeout_ms')::bigint,
701701
(p_config->'graph'->>'max_token_usage')::integer,
702702
ROW(
703-
p_config->'graph'->'model'->>'provider',
703+
COALESCE(p_config->'graph'->'model'->>'model_provider', p_config->'graph'->'model'->>'provider'),
704704
p_config->'graph'->'model'->>'model_name',
705705
(p_config->'graph'->'model'->>'temperature')::numeric(3,2),
706706
(p_config->'graph'->'model'->>'max_tokens')::integer

packages/database/src/queries/agents/queries.ts

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -777,7 +777,7 @@ export namespace agents {
777777
): Promise<ModelConfig | null> {
778778
const query = new Postgres.Query(
779779
`SELECT
780-
(model).model_provider as "provider",
780+
(model).model_provider as "model_provider",
781781
(model).model_name as "model_name",
782782
(model).temperature as "temperature",
783783
(model).max_tokens as "max_tokens"

packages/server/src/agents.storage.ts

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -366,7 +366,7 @@ export class AgentStorage implements OnModuleInit {
366366

367367
// const model = await this.getModelFromUser(userId); // Need to be used when user_id table will be created
368368
const model: ModelConfig = {
369-
provider: process.env.DEFAULT_MODEL_PROVIDER as string,
369+
model_provider: process.env.DEFAULT_MODEL_PROVIDER as string,
370370
model_name: process.env.DEFAULT_MODEL_NAME as string,
371371
temperature: parseFloat(process.env.DEFAULT_TEMPERATURE ?? '0.7'),
372372
max_tokens: parseInt(process.env.DEFAULT_MAX_TOKENS ?? '4096'),

0 commit comments

Comments
 (0)