Neum AI does for vector embeddings what Fivetran and Airbyte do for traditional data. But Neum AI goes beyond data loading and cleansing and is best-of-breed in optimizing the creation and real-time synchronization of vector embeddings at a massive scale. Our vision includes handling vector replication across data stores and embedding management for auditing and compliance. To get in touch with us send an email to founders@tryneum.com or book time in our Calendly: https://calendly.com/neum-ai/neum-ai-demo
David started his career working in gaming, communications and developer experience at Microsoft. He is passionate in building solutions that help democratize technology, especially within the developer community.
Originally from Venezuela, Kevin is a Software Engineer who spent 5+years at Microsoft building Data and AI platforms. Kevin left Microsoft to go work at Circle - the creators of USDC - building a service for ingesting and labeling blockchain data. Now, he is focused on building Neum AI, the best way to synchronize data sources with vector stores for companies building Generative AI applications. On his free time, you will find him cooking and hosting people over, his passion after programming.
TLDR: Neum AI Vector Sync (https://neum.ai) helps companies have accurate and up-to-date context in their LLM prompts. Connect and manage context from diverse data sources from Blob Storage to Notion to SQL and sync it into vector stores for searching.
Hey, David and Kevin here 👋, we are the cofounders of Neum AI!
From talking to dozens of developers building LLM applications, we know that building a prototype app is quite simple. Productizing that app can be challenging. Specifically, managing context for prompts can be tricky with data having to be synced from several different sources and maintained. With this, we decided to launch our managed vector sync platform in beta today!
While building Gen AI apps, we found that passing context to the prompt is crucial. However, the context must be accurate and up to date. If you are doing search over a vector database and that DB is not updated, then the search results are not good, and the quality of your context will suffer.
We found that people didn’t want to maintain syncing mechanisms between multiple data stores, especially for those dealing with lots of changing documents and source data. Here were some of the use cases:
With this, we set to create an easy-to-use UX and a set of APIs that help you manage and synchronize your data.
With Neum AI, your context in prompts is always accurate and up to date. It enables you to create a pipeline that keeps your data synced between a source (ex. Document DB) and a sink (ex. Pinecone). Think about FiveTran or Azure Data Factory for LLM data.
(We have APIs too 🥷💻)
As part of the synchronization process, Neum AI automatically embeds your data into vector form. This enables you to leverage common vector search techniques to do fast semantic look ups, ensuring that the context added to your prompt is accurate and helpful.
Neum AI is extensible to allow for you to connect a variety of data sources, bring your own embedding model and choose from different vector stores. Once you created the pipeline, we’ll handle the syncing, manageability and observability of it and alert you if something goes wrong. This ensures that your data is always fresh and your context in prompts is always accurate and up to date.
Once the pipeline is set, you can either run it once or specify a schedule for how often you want this pipeline to run, with of course, real time support in the works.
Sure! It is not an impossible task for a data engineer or developer who understands the process of extracting, embedding and loading data into a vector store. You would have to:
None of this is impossible, we simply take care of the mundane task of building it and making sure it is robust so that you focus on your business logic for your application.
Kevin and David, come from Venezuela and Costa Rica respectively. They met 10 years ago in college. You will find both of them building bots to automate tennis court reservations systems or coming up with new recipes.
David worked for 5+ years at Microsoft starting at Mixer and making his way to Azure. For the past several years, David focused on building developer experience for Azure Communication Services. Combining pro-code, low-code and no-code tools to help democratize adoption for developers.
Kevin worked at Microsoft for 6+ years building Big Data and AI platforms from the ground up. He worked in numerous hackathons that ended up in the hands of CVPs and EVPs. Additionally, he spent a year at Circle building API services for the core Data Engineering team working with Blockchain data at scale.
Combined they have a passion for building large-scale systems that are powerful, yet simple for developers to use. They were inspired by working with Azure Data Factory (Kevin being a contributor to the Microsoft Graph Data Connect project) to build an easy-to-configure, scalable, and secure data solution for companies using vector stores to build generative AI applications.
If you are building applications with semantic search or LLMs that leverage vector stores, come talk to us! (Book a time or email us) Head over to https://neum.ai to start using Neum AI today.