Wouldn't it be nice if we have language specific llms that work on average computers.
Like LLM that only trained on Python 3+, certain frameworks, certain code repos.
Then you can use a different model for searching the internet to implement different things to cut down on costs.
I used the regular home power stuff many years ago and the speeds were pretty bad and the network loss was unreliable.
My understanding is that it has improved in some circumstances, but if the connection ends up "hopping" through your breaker you get back to garbage speeds.
In theory you can get 2 Gbps speeds, but in practice it seems like still around 500 Mbps. I don't know if the loss has improved but it was a significant problem before, since even a low loss will render a connection unusable.
What happens if I want to make the video on the fly and save that to reuse it when the same question or topic comes up. No need to render a video. Just play the existing one.
This isn't natively supported -- we are continuously streaming frames throughout the conversation session that are generated in real-time. If you were building your own conversational AI pipeline (e.g. using our LiveKit integration), I suppose it would be possible to route things like this with your own logic. But it would probably include jump cuts and not look as good.
I just followed the Quickstart[1] in the GitHub repo, refreshingly straight forward. Using the pip package worked fine, as did installing the editable version using the git repository. Just install the CUDA version of PyTorch[2] first.
The HF demo is very similar to the GitHub demo, so easy to try out.
That's for CUDA 12.8, change PyTorch install accordingly.
Skipped FlashAttention since I'm on Windows and I haven't gotten FlashAttention 2 to work there yet (I found some precompiled FA3 files[3] but Qwen3-TTS isn't FA3 compatible yet).
Like LLM that only trained on Python 3+, certain frameworks, certain code repos. Then you can use a different model for searching the internet to implement different things to cut down on costs.
Maybe I have no idea what I'm talking about lol
reply