mirror of
https://github.com/NovaOSS/nova-python.git
synced 2024-11-25 22:43:58 +01:00
Compare commits
No commits in common. "1361d09025ee05fef1ffb5ed3186981d99eb4624" and "ccbaa4e247e070fc90323b1d4ba3d477d61147b0" have entirely different histories.
1361d09025
...
ccbaa4e247
|
@ -1,6 +1,6 @@
|
|||
[package]
|
||||
name = "nova-python"
|
||||
version = "0.1.7"
|
||||
version = "0.1.6"
|
||||
edition = "2021"
|
||||
|
||||
# See more keys and their definitions at https://doc.rust-lang.org/cargo/reference/manifest.html
|
||||
|
@ -9,7 +9,7 @@ name = "nova_python"
|
|||
crate-type = ["cdylib"]
|
||||
|
||||
[dependencies]
|
||||
pyo3 = {version = "0.19.2", features = ["generate-import-lib"]}
|
||||
pyo3 = {version = "0.19.2", features = ["extension-module", "generate-import-lib"]}
|
||||
reqwest = "0.11.18"
|
||||
tokio = { version = "1.29.1", features = ["rt-multi-thread", "time"] }
|
||||
serde_json = "1.0.104"
|
12
README.md
12
README.md
|
@ -36,7 +36,7 @@ Now, to make a request, use the `make_request` function. For example:
|
|||
from nova_python import Endpoints, Models, NovaClient
|
||||
client = NovaClient("YOUR_API_KEY")
|
||||
|
||||
response = client.make_request(
|
||||
reponse = client.make_request(
|
||||
endpoint=Endpoints.CHAT_COMPLETION,
|
||||
model=Models.GPT3,
|
||||
data=[
|
||||
|
@ -54,7 +54,7 @@ or
|
|||
from nova_python import Endpoints, Models, NovaClient
|
||||
client = NovaClient("YOUR_API_KEY")
|
||||
|
||||
response = client.make_request(
|
||||
reponse = client.make_request(
|
||||
endpoint=Endpoints.MODERATION,
|
||||
model=Models.MODERATION_STABLE,
|
||||
data=[{"input": "I'm going to kill them."}]
|
||||
|
@ -65,20 +65,20 @@ response = client.make_request(
|
|||
If everything goes to plan, you'll receive a string containing JSON-Data, which you can then use in your project.
|
||||
|
||||
Note, that when using chat completion, as special ChatResponse-Instance get's returned.
|
||||
You can access the response's json.data, by casting it to a string using the `str` method, like this:
|
||||
You can access the reponse's json.data, by casting it to a string using the `str` method, like this:
|
||||
```python
|
||||
...
|
||||
|
||||
str(response)
|
||||
str(reponse)
|
||||
|
||||
...
|
||||
```
|
||||
|
||||
but more importantly, you can use it's `get_message_content` function, to directly get access to the chat response. Used like this:
|
||||
but more importantly, you can use it's `get_message_content` function, to directly get access to the chat reponse. Used like this:
|
||||
```
|
||||
...
|
||||
|
||||
content = response.get_message_content()
|
||||
content = reponse.get_message_content()
|
||||
|
||||
...
|
||||
```
|
||||
|
|
23
src/lib.rs
23
src/lib.rs
|
@ -231,14 +231,23 @@ impl ChatResponse {
|
|||
}
|
||||
|
||||
fn model_is_compatible(endpoint: &Endpoints, model: &Models) -> bool {
|
||||
let chat_models = [Models::Gpt3, Models::Gpt4];
|
||||
let moderation_models = [Models::ModerationStable, Models::ModerationLatest];
|
||||
|
||||
match endpoint {
|
||||
Endpoints::ChatCompletion => chat_models.contains(model),
|
||||
Endpoints::Moderation => moderation_models.contains(model),
|
||||
_ => false
|
||||
if endpoint == &Endpoints::ChatCompletion {
|
||||
if [Models::Gpt3, Models::Gpt4].contains(model) {
|
||||
return true;
|
||||
} else {
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
||||
else if endpoint == &Endpoints::Moderation {
|
||||
if [Models::ModerationStable, Models::ModerationLatest].contains(model) {
|
||||
return true;
|
||||
} else {
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
||||
false
|
||||
}
|
||||
|
||||
fn key_is_valid(api_key: &str) -> bool {
|
||||
|
|
Loading…
Reference in a new issue