Posts Tagged pinecone

Deploy Low-Code LLM Apps on AWS with Flowise and LocalAI

Deploy Low-Code LLM Apps on AWS with Flowise and LocalAI

Building low-code LLM applications just got easier! With Flowise, LocalAI, and Pulumi, you can quickly create, deploy, and manage AI-powered workflows on AWS EKS—without writing complex code.

In this guide, you’ll learn how to:

  1. Set up a low-code LLM application using Flowise & LocalAI
  2. Deploy it on AWS EKS with Pulumi & TypeScript
  3. Create a chatbot workflow and expose it via API

Flowise simplifies LLM orchestration with a drag-and-drop interface, while LocalAI enables local inference without requiring a GPU. Pulumi streamlines cloud infrastructure deployment with just a few lines of code.

Read more →

Pinecone Provider Now Available for Pulumi

Pinecone Provider Now Available for Pulumi

Hello, Pulumi Pinecone Provider! 👋

The Pinecone integration with Pulumi offers a native way to manage Pinecone indexes, including the newly-announced serverless indexes. Utilize any of Pulumi’s supported languages to effortlessly create, update, and remove your Pinecone indexes. This integration facilitates the application of Infrastructure as Code principles, helping you to work even more efficiently. Furthermore, this gives you the benefit of tapping into Pulumi’s wide range of providers, offering you a diverse and powerful set of tools to enhance your development work.

Read more →