Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

upgrade to latest cartesia 1.0.3 #587

Merged
merged 6 commits into from
Jul 3, 2024

Conversation

rjheeta
Copy link
Contributor

@rjheeta rjheeta commented Jun 27, 2024

Upgrade to latest Cartesia

Copy link
Contributor

@ajar98 ajar98 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

happy to get this in without the streaming update as well! though this will make the synthesizer much faster :)

pyproject.toml Outdated
@@ -43,7 +43,7 @@ groq = { version = "^0.9.0", optional = true }
# Synthesizers
google-cloud-texttospeech = { version = "^2.16.3", optional = true }
pvkoala = { version = "^2.0.1", optional = true }
cartesia = { version = "^0.1.1", optional = true }
cartesia = "^1.0.3"
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

can this go back to being optional?



async def create_speech_uncached(
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

if interested, would love to support you making this properly streaming!!

would be something like:

def chunk_generator(sse):
  async for chunk in sse:
    yield SynthesisResult.ChunkResult(data=chunk)
    
return SynthesisResult(chunk_generator=chunk_generator(generator), lambda seconds: self.get_message_cutoff_from_voice_speed(message, seconds, 150))

you would have to also make the synthesizer support mulaw properly (which I think they fixed recently!)

@ajar98 ajar98 merged commit 949711f into vocodedev:main Jul 3, 2024
4 checks passed
ajar98 added a commit that referenced this pull request Jul 5, 2024
* [DOW-118] set up code linting and tests (#589)

* adds github workflow

* run black

* run isort

* adds precommit

* adds vscode settings

* adds pre-commit guidelines (#590)

* creates docker image, updates telephony app deps (#601)

* [DOW-105] refactor interruptions into the output device (#586)

* [DOW-105] refactor interruptions into the output device (#562)

* initial refactor works

* remove notion of UtteranceAudioChunk and put all of the state in the callback

* move per_chunk_allowance_seconds into output device

* onboard onto vonage

* rename to abstract output device and onboard other output devices

* initial work to onboard twilio output device

* twilio conversation works

* some cleanup with better comments

* unset poetry.lock

* move abstract play method into ratelimitoutputdevice + dispatch to thread in fileoutputdevice

* rename back to AsyncWorker

* comments

* work through a bit of mypy

* asyncio.gather is g2g:

* create interrupt lock

* remove todo

* remove last todo

* remove log for interrupts

* fmt

* fix mypy

* fix mypy

* isort

* creates first test and adds scaffolding

* adds two other send_speech_to_output tests

* make send_speech_to_output more efficient

* adds tests for rate limit interruptions output device

* makes some variables private and also makes the chunk id coming back from the mark match the incoming audio chunk

* adds twilio output device tests

* make typing better for output devices

* fix mypy

* resolve PR comments

* resolve PR comments

* [DOW-101] LiveKit integration (#591)

* checkpoint

* livekit v0

* in progress changes

* integrate with worker

* fix import

* update deps and remove unneeded files

* integrate it properly into app

* fix interrupts

* make transcript publish work

* a confounding fix

* isort

* constants, some cleanup

---------

Co-authored-by: Kian <[email protected]>

* upgrade to latest cartesia 1.0.3 (#587)

* upgrade to latest cartesia 1.0.3

* fixed linting conflict

* finish streaming

* make cartesia optional

---------

Co-authored-by: Ajay Raj <[email protected]>

* poetry version prerelease (#602)

* feat: Add ability to configure OpenAI base URL in ChatGPTAgentConfig (#577)

* feat: Add ability to configure OpenAI base URL in ChatGPTAgentConfig

- Added `base_url` parameter to `ChatGPTAgentConfig` to allow customization of the OpenAI API base URL.
- Updated `instantiate_openai_client` function to use the `base_url` parameter from the configuration.
- Modified `ChatGPTAgent` to utilize the updated `instantiate_openai_client` function.
- Added tests to verify the new `base_url` functionality in `tests/streaming/agent/test_base_agent.py`.

This enhancement allows users to specify a custom OpenAI API base URL, providing greater flexibility in agent configuration.

* adding capability to use the openai compatible endpoint with token estimation for llama

* lint fix

* changing openai base_url parameter for overall less code changes

* missed logging update

* Update vocode/streaming/agent/chat_gpt_agent.py

* Update tests/streaming/agent/test_base_agent.py

* fix test

---------

Co-authored-by: Ajay Raj <[email protected]>

* Support passthrough of AsyncHTTPTransport (#603)

Support passthrough of AsyncHTTPTransport object

* add script used to make PR

* adds test target for vocodehq-public

* Remove catch-all exception logger for asyncio tasks (#605)

* remove error log from exception for asyncio tasks

* remove log error on chatgpt query

---------

Co-authored-by: Kian <[email protected]>
Co-authored-by: rjheeta <[email protected]>
Co-authored-by: Clay Elmore <[email protected]>
Co-authored-by: vocode-petern <[email protected]>
Co-authored-by: Adnaan Sachidanandan <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants