Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feat: Remote API Parameters Correction #2802

Merged
merged 44 commits into from
May 4, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
44 commits
Select commit Hold shift + click to select a range
9aa6d0e
fix: change to gpt4 turbo
hahuyhoang411 Apr 24, 2024
973e9de
add: params
hahuyhoang411 Apr 24, 2024
c355137
fix: change to gpt 3.5 turbo
hahuyhoang411 Apr 24, 2024
fd10ef4
delete: redundant
hahuyhoang411 Apr 24, 2024
6785e1d
fix: correct description
hahuyhoang411 Apr 24, 2024
eb9054c
version bump
hahuyhoang411 Apr 24, 2024
10d1672
add: params
hahuyhoang411 Apr 24, 2024
e18fd07
fix: version bump
hahuyhoang411 Apr 24, 2024
90d30b3
delete: deprecated
hahuyhoang411 Apr 24, 2024
f4c10c9
add: params
hahuyhoang411 Apr 24, 2024
1fd75a1
add: new model
hahuyhoang411 Apr 24, 2024
d794a19
chore: version bump
hahuyhoang411 Apr 24, 2024
0a64278
fix: version correct
hahuyhoang411 Apr 24, 2024
5e8eaf5
add: params
hahuyhoang411 Apr 24, 2024
9a87063
fix: version bump
hahuyhoang411 Apr 24, 2024
6aad2e5
fix: change to gpt4 turbo
hahuyhoang411 Apr 24, 2024
1344c55
add: params
hahuyhoang411 Apr 24, 2024
f219404
fix: change to gpt 3.5 turbo
hahuyhoang411 Apr 24, 2024
222702d
delete: redundant
hahuyhoang411 Apr 24, 2024
df2ec56
fix: correct description
hahuyhoang411 Apr 24, 2024
fb13cf3
version bump
hahuyhoang411 Apr 24, 2024
e8e92d6
add: params
hahuyhoang411 Apr 24, 2024
3f5faa7
fix: version bump
hahuyhoang411 Apr 24, 2024
806d77a
delete: deprecated
hahuyhoang411 Apr 24, 2024
b71242a
add: params
hahuyhoang411 Apr 24, 2024
adfafd5
add: new model
hahuyhoang411 Apr 24, 2024
d454c08
chore: version bump
hahuyhoang411 Apr 24, 2024
12400f2
fix: version correct
hahuyhoang411 Apr 24, 2024
5337669
add: params
hahuyhoang411 Apr 24, 2024
4b0e215
fix: version bump
hahuyhoang411 Apr 24, 2024
e935392
fix: llama2 no longer supported
hahuyhoang411 Apr 25, 2024
942904a
fix: reverse mistral api
hahuyhoang411 May 2, 2024
8b9f8c6
fix: conflicts
hahuyhoang411 May 2, 2024
4610d4d
fix: add params
hahuyhoang411 May 2, 2024
241e6a8
Merge branch 'dev' into chore/remote-api-params
hahuyhoang411 May 2, 2024
2e68b85
fix: mistral api redundant params
hahuyhoang411 May 2, 2024
489c33a
Merge branch 'chore/remote-api-params' of https://github.com/janhq/ja…
hahuyhoang411 May 2, 2024
a9e556d
fix: typo
hahuyhoang411 May 2, 2024
26b81d9
fix: typo
hahuyhoang411 May 2, 2024
21e84d6
fix: correct context length
hahuyhoang411 May 3, 2024
46b2234
Merge branch 'chore/remote-api-params' of https://github.com/janhq/ja…
hahuyhoang411 May 3, 2024
29b8f0d
Merge branch 'dev' into chore/remote-api-params
Van-QA May 3, 2024
0ff0065
fix: remove stop
hahuyhoang411 May 3, 2024
4e94703
Merge branch 'chore/remote-api-params' of https://github.com/janhq/ja…
hahuyhoang411 May 3, 2024
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
2 changes: 1 addition & 1 deletion extensions/inference-groq-extension/package.json
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
{
"name": "@janhq/inference-groq-extension",
"productName": "Groq Inference Engine",
"version": "1.0.0",
"version": "1.0.1",
"description": "This extension enables fast Groq chat completion API calls",
"main": "dist/index.js",
"module": "dist/module.js",
Expand Down
88 changes: 30 additions & 58 deletions extensions/inference-groq-extension/resources/models.json
Original file line number Diff line number Diff line change
Expand Up @@ -8,18 +8,18 @@
"id": "llama3-70b-8192",
"object": "model",
"name": "Groq Llama 3 70b",
"version": "1.0",
"version": "1.1",
"description": "Groq Llama 3 70b with supercharged speed!",
"format": "api",
"settings": {
"text_model": false
},
"settings": {},
"parameters": {
"max_tokens": 8192,
"temperature": 0.7,
"top_p": 1,
"stop": null,
"stream": true
"top_p": 0.95,
"stream": true,
"stop": [],
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "Meta",
Expand All @@ -36,18 +36,18 @@
"id": "llama3-8b-8192",
"object": "model",
"name": "Groq Llama 3 8b",
"version": "1.0",
"version": "1.1",
"description": "Groq Llama 3 8b with supercharged speed!",
"format": "api",
"settings": {
"text_model": false
},
"settings": {},
"parameters": {
"max_tokens": 8192,
"temperature": 0.7,
"top_p": 1,
"stop": null,
"stream": true
"top_p": 0.95,
"stream": true,
"stop": [],
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "Meta",
Expand All @@ -64,53 +64,25 @@
"id": "gemma-7b-it",
"object": "model",
"name": "Groq Gemma 7b Instruct",
"version": "1.0",
"version": "1.1",
"description": "Groq Gemma 7b Instruct with supercharged speed!",
"format": "api",
"settings": {
"text_model": false
},
"settings": {},
"parameters": {
"max_tokens": 4096,
"max_tokens": 8192,
"temperature": 0.7,
"top_p": 1,
"stop": null,
"stream": true
"top_p": 0.95,
"stream": true,
"stop": [],
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "Google",
"tags": ["General"]
},
"engine": "groq"
},
{
"sources": [
{
"url": "https://groq.com"
}
],
"id": "llama2-70b-4096",
"object": "model",
"name": "Groq Llama 2 70b",
"version": "1.0",
"description": "Groq Llama 2 70b with supercharged speed!",
"format": "api",
"settings": {
"text_model": false
},
"parameters": {
"max_tokens": 4096,
"temperature": 0.7,
"top_p": 1,
"stop": null,
"stream": true
},
"metadata": {
"author": "Meta",
"tags": ["General", "Big Context Length"]
},
"engine": "groq"
},
{
"sources": [
{
Expand All @@ -120,18 +92,18 @@
"id": "mixtral-8x7b-32768",
"object": "model",
"name": "Groq Mixtral 8x7b Instruct",
"version": "1.0",
"version": "1.1",
"description": "Groq Mixtral 8x7b Instruct is Mixtral with supercharged speed!",
"format": "api",
"settings": {
"text_model": false
},
"settings": {},
"parameters": {
"max_tokens": 4096,
"max_tokens": 32768,
"temperature": 0.7,
"top_p": 1,
"stop": null,
"stream": true
"top_p": 0.95,
"stream": true,
"stop": [],
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "Mistral",
Expand Down
2 changes: 1 addition & 1 deletion extensions/inference-mistral-extension/package.json
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
{
"name": "@janhq/inference-mistral-extension",
"productName": "MistralAI Inference Engine",
"version": "1.0.0",
"version": "1.0.1",
"description": "This extension enables Mistral chat completion API calls",
"main": "dist/index.js",
"module": "dist/module.js",
Expand Down
52 changes: 25 additions & 27 deletions extensions/inference-mistral-extension/resources/models.json
Original file line number Diff line number Diff line change
Expand Up @@ -8,20 +8,20 @@
"id": "mistral-small-latest",
"object": "model",
"name": "Mistral Small",
"version": "1.0",
"description": "Mistral Small is the ideal choice for simpe tasks that one can do in builk - like Classification, Customer Support, or Text Generation. It offers excellent performance at an affordable price point.",
"version": "1.1",
"description": "Mistral Small is the ideal choice for simple tasks (Classification, Customer Support, or Text Generation) at an affordable price.",
"format": "api",
"settings": {},
"parameters": {
"max_tokens": 4096,
"temperature": 0.7
"max_tokens": 32000,
"temperature": 0.7,
"top_p": 0.95,
"stream": true
},
"metadata": {
"author": "Mistral",
"tags": [
"Classification",
"Customer Support",
"Text Generation"
"General"
]
},
"engine": "mistral"
Expand All @@ -32,24 +32,23 @@
"url": "https://docs.mistral.ai/api/"
}
],
"id": "mistral-medium-latest",
"id": "mistral-large-latest",
"object": "model",
"name": "Mistral Medium",
"version": "1.0",
"description": "Mistral Medium is the ideal for intermediate tasks that require moderate reasoning - like Data extraction, Summarizing a Document, Writing a Job Description, or Writing Product Descriptions. Mistral Medium strikes a balance between performance and capability, making it suitable for a wide range of tasks that only require language transformaion",
"name": "Mistral Large",
"version": "1.1",
"description": "Mistral Large is ideal for complex tasks (Synthetic Text Generation, Code Generation, RAG, or Agents).",
"format": "api",
"settings": {},
"parameters": {
"max_tokens": 4096,
"temperature": 0.7
"max_tokens": 32000,
"temperature": 0.7,
"top_p": 0.95,
"stream": true
},
"metadata": {
"author": "Mistral",
"tags": [
"Data extraction",
"Summarizing a Document",
"Writing a Job Description",
"Writing Product Descriptions"
"General"
]
},
"engine": "mistral"
Expand All @@ -60,24 +59,23 @@
"url": "https://docs.mistral.ai/api/"
}
],
"id": "mistral-large-latest",
"id": "open-mixtral-8x22b",
"object": "model",
"name": "Mistral Large",
"version": "1.0",
"description": "Mistral Large is ideal for complex tasks that require large reasoning capabilities or are highly specialized - like Synthetic Text Generation, Code Generation, RAG, or Agents.",
"name": "Mixtral 8x22B",
"version": "1.1",
"description": "Mixtral 8x22B is a high-performance, cost-effective model designed for complex tasks.",
"format": "api",
"settings": {},
"parameters": {
"max_tokens": 4096,
"temperature": 0.7
"max_tokens": 32000,
"temperature": 0.7,
"top_p": 0.95,
"stream": true
},
"metadata": {
"author": "Mistral",
"tags": [
"Text Generation",
"Code Generation",
"RAG",
"Agents"
"General"
]
},
"engine": "mistral"
Expand Down
2 changes: 1 addition & 1 deletion extensions/inference-openai-extension/package.json
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
{
"name": "@janhq/inference-openai-extension",
"productName": "OpenAI Inference Engine",
"version": "1.0.0",
"version": "1.0.1",
"description": "This extension enables OpenAI chat completion API calls",
"main": "dist/index.js",
"module": "dist/module.js",
Expand Down
59 changes: 24 additions & 35 deletions extensions/inference-openai-extension/resources/models.json
Original file line number Diff line number Diff line change
Expand Up @@ -5,20 +5,25 @@
"url": "https://openai.com"
}
],
"id": "gpt-4",
"id": "gpt-4-turbo",
"object": "model",
"name": "OpenAI GPT 4",
"version": "1.0",
"version": "1.1",
"description": "OpenAI GPT 4 model is extremely good",
"format": "api",
"settings": {},
"parameters": {
"max_tokens": 4096,
"temperature": 0.7
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"stop": [],
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "OpenAI",
"tags": ["General", "Big Context Length"]
"tags": ["General"]
},
"engine": "openai"
},
Expand All @@ -31,43 +36,22 @@
"id": "gpt-4-vision-preview",
"object": "model",
"name": "OpenAI GPT 4 with Vision (Preview)",
"version": "1.0",
"description": "OpenAI GPT 4 with Vision model is extremely good in preview",
"version": "1.1",
"description": "OpenAI GPT-4 Vision model features vision understanding capabilities",
"format": "api",
"settings": {
"vision_model": true,
"textModel": false
},
"parameters": {
"max_tokens": 4096,
"temperature": 0.7
"temperature": 0.7,
"top_p": 0.95,
"stream": true
},
"metadata": {
"author": "OpenAI",
"tags": ["General", "Big Context Length", "Vision"]
},
"engine": "openai"
},
{
"sources": [
{
"url": "https://openai.com"
}
],
"id": "gpt-3.5-turbo-16k-0613",
"object": "model",
"name": "OpenAI GPT 3.5 Turbo 16k 0613",
"version": "1.0",
"description": "OpenAI GPT 3.5 Turbo 16k 0613 model is extremely good",
"format": "api",
"settings": {},
"parameters": {
"max_tokens": 4096,
"temperature": 0.7
},
"metadata": {
"author": "OpenAI",
"tags": ["General", "Big Context Length"]
"tags": ["General", "Vision"]
},
"engine": "openai"
},
Expand All @@ -80,17 +64,22 @@
"id": "gpt-3.5-turbo",
"object": "model",
"name": "OpenAI GPT 3.5 Turbo",
"version": "1.0",
"description": "OpenAI GPT 3.5 Turbo model is extremely good",
"version": "1.1",
"description": "OpenAI GPT 3.5 Turbo model is extremely fast",
"format": "api",
"settings": {},
"parameters": {
"max_tokens": 4096,
"temperature": 0.7
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"stop": [],
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "OpenAI",
"tags": ["General", "Big Context Length"]
"tags": ["General"]
},
"engine": "openai"
}
Expand Down