@@ -11,14 +11,10 @@ def __init__(self, client=None, config=None):
11
11
12
12
# default parameters - can be overrided using config
13
13
self .temperature = 0.7
14
- self .max_tokens = 500
15
14
16
15
if "temperature" in config :
17
16
self .temperature = config ["temperature" ]
18
17
19
- if "max_tokens" in config :
20
- self .max_tokens = config ["max_tokens" ]
21
-
22
18
if "api_type" in config :
23
19
raise Exception (
24
20
"Passing api_type is now deprecated. Please pass an OpenAI client instead."
@@ -75,7 +71,6 @@ def submit_prompt(self, prompt, **kwargs) -> str:
75
71
response = self .client .chat .completions .create (
76
72
model = model ,
77
73
messages = prompt ,
78
- max_tokens = self .max_tokens ,
79
74
stop = None ,
80
75
temperature = self .temperature ,
81
76
)
@@ -87,7 +82,6 @@ def submit_prompt(self, prompt, **kwargs) -> str:
87
82
response = self .client .chat .completions .create (
88
83
engine = engine ,
89
84
messages = prompt ,
90
- max_tokens = self .max_tokens ,
91
85
stop = None ,
92
86
temperature = self .temperature ,
93
87
)
@@ -98,7 +92,6 @@ def submit_prompt(self, prompt, **kwargs) -> str:
98
92
response = self .client .chat .completions .create (
99
93
engine = self .config ["engine" ],
100
94
messages = prompt ,
101
- max_tokens = self .max_tokens ,
102
95
stop = None ,
103
96
temperature = self .temperature ,
104
97
)
@@ -109,7 +102,6 @@ def submit_prompt(self, prompt, **kwargs) -> str:
109
102
response = self .client .chat .completions .create (
110
103
model = self .config ["model" ],
111
104
messages = prompt ,
112
- max_tokens = self .max_tokens ,
113
105
stop = None ,
114
106
temperature = self .temperature ,
115
107
)
@@ -123,7 +115,6 @@ def submit_prompt(self, prompt, **kwargs) -> str:
123
115
response = self .client .chat .completions .create (
124
116
model = model ,
125
117
messages = prompt ,
126
- max_tokens = self .max_tokens ,
127
118
stop = None ,
128
119
temperature = self .temperature ,
129
120
)
0 commit comments