I've been experimenting with TypingMind, a web-based wrapper for various LLMs, and recreated parts of my Claude setup with third-party apps at a much higher speed.
I started playing around with TypingMind, hosted on Grow, running with a custom plugin, and achieved fast inference with Kimi K2.
I’ll talk about this more in depth in Monday’s episode of AppStories.
A video demonstration shows Kimi K2 hosted on Groq on the left.
Author's summary: Experimenting with TypingMind and Kimi K2 for fast AI inference.