Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -271,7 +271,7 @@ class VideoAnalytics:
|
|
| 271 |
]
|
| 272 |
# Generate completion using ChatGPT model
|
| 273 |
response = self.client.chat.completions.create(
|
| 274 |
-
model="
|
| 275 |
messages=conversation,
|
| 276 |
temperature=0,
|
| 277 |
max_tokens=1000
|
|
@@ -315,7 +315,7 @@ class VideoAnalytics:
|
|
| 315 |
]
|
| 316 |
# Generate completion using ChatGPT model
|
| 317 |
response = self.client.chat.completions.create(
|
| 318 |
-
model="
|
| 319 |
messages=conversation,
|
| 320 |
temperature=0,
|
| 321 |
max_tokens=1000
|
|
@@ -355,7 +355,7 @@ class VideoAnalytics:
|
|
| 355 |
]
|
| 356 |
# Generate completion using ChatGPT model
|
| 357 |
response = self.client.chat.completions.create(
|
| 358 |
-
model="
|
| 359 |
messages=conversation,
|
| 360 |
temperature=0,
|
| 361 |
max_tokens=1000
|
|
@@ -394,7 +394,7 @@ class VideoAnalytics:
|
|
| 394 |
]
|
| 395 |
# Generate completion using ChatGPT model
|
| 396 |
response = self.client.chat.completions.create(
|
| 397 |
-
model="
|
| 398 |
messages=conversation,
|
| 399 |
temperature=0,
|
| 400 |
max_tokens=1000
|
|
|
|
| 271 |
]
|
| 272 |
# Generate completion using ChatGPT model
|
| 273 |
response = self.client.chat.completions.create(
|
| 274 |
+
model="GPT-3",
|
| 275 |
messages=conversation,
|
| 276 |
temperature=0,
|
| 277 |
max_tokens=1000
|
|
|
|
| 315 |
]
|
| 316 |
# Generate completion using ChatGPT model
|
| 317 |
response = self.client.chat.completions.create(
|
| 318 |
+
model="GPT-3",
|
| 319 |
messages=conversation,
|
| 320 |
temperature=0,
|
| 321 |
max_tokens=1000
|
|
|
|
| 355 |
]
|
| 356 |
# Generate completion using ChatGPT model
|
| 357 |
response = self.client.chat.completions.create(
|
| 358 |
+
model="GPT-3",
|
| 359 |
messages=conversation,
|
| 360 |
temperature=0,
|
| 361 |
max_tokens=1000
|
|
|
|
| 394 |
]
|
| 395 |
# Generate completion using ChatGPT model
|
| 396 |
response = self.client.chat.completions.create(
|
| 397 |
+
model="GPT-3",
|
| 398 |
messages=conversation,
|
| 399 |
temperature=0,
|
| 400 |
max_tokens=1000
|