<oembed><type>rich</type><version>1.0</version><author_name>hoshi (npub185…4x93y)</author_name><author_url>https://nostr.ae/npub185cl0jg373tl9hqaa000z0znzw7r5c909nkc2ndf9mvmgenqxpjqf4x93y</author_url><provider_name>njump</provider_name><provider_url>https://nostr.ae</provider_url><html>I get similar responses when I run a quantized LLM locally on my phone</html></oembed>