Event JSON
{
"id": "a59536db03322e87a2681149d6f5f64824db75290f7ae1aacd90ecb988a34aa5",
"pubkey": "efd4d4a38bcd0004d31031a17972a6a6a3b32fe6f3953958c47d6325f2b8d106",
"created_at": 1721141794,
"kind": 1,
"tags": [
[
"t",
"PromptQL"
],
[
"t",
"texttospeech"
],
[
"t",
"llm"
],
[
"t",
"locallm"
],
[
"t",
"promptengineering"
],
[
"t",
"prompts"
],
[
"t",
"messageprotocol"
],
[
"t",
"agents"
],
[
"t",
"golang"
],
[
"t",
"go"
],
[
"t",
"programming"
],
[
"t",
"opensource"
],
[
"t",
"library"
],
[
"t",
"indietech"
],
[
"t",
"programming"
],
[
"t",
"ai"
],
[
"t",
"generativeai"
],
[
"t",
"blog"
]
],
"content": "4 months ago I tested my audio stories generator on the Nous Hermes 10.7B model which runs locally on my CPU. And on llama.cpp server. And there are results! \n\nThe stories generator (built with PromptQL front-end and Nous Hermes, OpenAI TTS back-ends) is here:\n\nhttps://gitlab.com/jbyte777/storybox-generator/-/tree/with-nous-hermes-10.7b\n \nThe Nous Hermes 10.7B model can be downloaded here:\n\nhttps://huggingface.co/TheBloke/Nous-Hermes-2-SOLAR-10.7B-GGUF \n\nThe llama.cpp server can be installed from here:\n\nhttps://github.com/ggerganov/llama.cpp/tree/master/examples/server \n\n#PromptQL #texttospeech #llm #locallm #promptengineering #prompts #messageprotocol #agents \n#golang #go #programming #opensource #library #indietech #programming #ai #generativeai \n#blog",
"sig": "f6b2486c8dcc88d5634e59c8db8c2cfc289fbc9f16828e24c94191dff8ce76f8d468539e31aef5db2c126c88ff71c249282e8ae97da9b05cf165d0808d77f507"
}