Update README.md

This commit is contained in:
Ajay Raj 2023-03-29 14:18:24 -07:00 committed by GitHub
commit a70528add5
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23

View file

@ -36,9 +36,15 @@ Vocode is an open source library that makes it easy to build voice-based LLM app
Check out our React SDK [here](https://github.com/vocodedev/vocode-react-sdk)! Check out our React SDK [here](https://github.com/vocodedev/vocode-react-sdk)!
# Contributing # 🫂 Contribution
If there are features or integrations that don't exist yet, please add them! Feel free to fork and create a PR and we will get it merged as soon as possible. We'll have more guidelines on contributions soon :) We'd love for you all to build on top of our abstractions to enable new and better LLM voice applications!
You can extend our [`BaseAgent`](https://github.com/vocodedev/vocode-python/blob/main/vocode/streaming/agent/base_agent.py), [`BaseTranscriber`](https://github.com/vocodedev/vocode-python/blob/main/vocode/streaming/transcriber/base_transcriber.py), and [`BaseSynthesizer`](https://github.com/vocodedev/vocode-python/blob/main/vocode/streaming/synthesizer/base_synthesizer.py) abstractions to integrate with new LLM APIs, speech recognition and speech synthesis providers. More detail [here](https://docs.vocode.dev/create-your-own-agent#self-hosted).
You can also work with our [`BaseInputDevice`](https://github.com/vocodedev/vocode-python/blob/main/vocode/streaming/input_device/base_input_device.py) and [`BaseOutputDevice`](https://github.com/vocodedev/vocode-python/blob/main/vocode/streaming/output_device/base_output_device.py) abstractions to set up voice applications on new surfaces/platforms. More guides for this coming soon!
Because our [`StreamingConversation`](https://github.com/vocodedev/vocode-python/blob/main/vocode/streaming/streaming_conversation.py) runs locally, it's relatively quick to develop on! Feel free to fork and create a PR and we will get it merged as soon as possible. And we'd love to talk to you on [Discord](https://discord.gg/NaU4mMgcnC)!
# 🚀 Quickstart (Self-hosted) # 🚀 Quickstart (Self-hosted)
@ -169,13 +175,3 @@ if __name__ == "__main__":
# 🌱 Documentation # 🌱 Documentation
[docs.vocode.dev](https://docs.vocode.dev/) [docs.vocode.dev](https://docs.vocode.dev/)
# Contribution
We'd love for you all to build on top of our abstractions to enable new and better LLM voice applications!
You can extend our [`BaseAgent`](https://github.com/vocodedev/vocode-python/blob/main/vocode/streaming/agent/base_agent.py), [`BaseTranscriber`](https://github.com/vocodedev/vocode-python/blob/main/vocode/streaming/transcriber/base_transcriber.py), and [`BaseSynthesizer`](https://github.com/vocodedev/vocode-python/blob/main/vocode/streaming/synthesizer/base_synthesizer.py) abstractions to integrate with new LLM APIs, speech recognition and speech synthesis providers. More detail [here](https://docs.vocode.dev/create-your-own-agent#self-hosted).
You can also work with our [`BaseInputDevice`](https://github.com/vocodedev/vocode-python/blob/main/vocode/streaming/input_device/base_input_device.py) and [`BaseOutputDevice`](https://github.com/vocodedev/vocode-python/blob/main/vocode/streaming/output_device/base_output_device.py) abstractions to set up voice applications on new surfaces/platforms. More guides for this coming soon!
Because our `StreamingConversation` takes in these objects and runs locally, it's relatively quick to develop on! We appreciate pull requests and would love to talk to your on [Community](https://discord.gg/NaU4mMgcnC).