Event JSON
{
"id": "c3720c5ececd9f4c97f3a13a70617c19ddaf2e87cc49f8978919927d3c1af171",
"pubkey": "77620d8f02483272621d387ad54c1bf612b11a9261c7f3927b408434eca7de3f",
"created_at": 1692211268,
"kind": 1,
"tags": [
[
"e",
"c3c8604e7b9073eac9f6e161bf4ec8b963297f6d901ac5fe4223d44a1c3d2d3a",
""
],
[
"e",
"a8c1b2b0af60da6387a680bf262926c90f230f25ac319f156295ccfd7ebb87a2"
],
[
"p",
"ea2e3c814d08a378f8a5b8faecb2884d05855975c5ca4b5c25e2d6f936286f14"
],
[
"p",
"5d9862e0dbe6b167dbf07c5a0496ecba773c348f7caaa0b3a07f52fe6dd1095f"
]
],
"content": "Currently it uses the Nous Hermes Llama 2 (7B). If you’re technical, you can customize the Dockerfile to run a different Llama model: https://github.com/getumbrel/llama-gpt/blob/c76225a6fc26a000fc07b074223a69b0d65b7bcf/api/Dockerfile#L6",
"sig": "876d43ae1edd35559238bb337903c40e878e626e337cb631bb1f67bec5c7d57ad7d37fab2db2f23713a384aef8243f64e462c0540e434fabf69ad4a66055c7dc"
}