Docker introduced a brand new GenAI Stack in partnership with Neo4j, LangChain, and Ollama throughout its annual DockerCon developer convention keynote. This GenAI Stack is designed to assist builders rapidly and simply construct generative AI functions with out trying to find and configuring numerous applied sciences.
It consists of pre-configured parts like massive language fashions (LLMs) from Ollama, vector and graph databases from Neo4j, and the LangChain framework. Docker additionally launched its first AI-powered product, Docker AI.
The GenAI Stack addresses fashionable use instances for generative AI and is on the market within the Docker Studying Middle and on GitHub. It provides pre-configured open-source LLMs, help from Ollama for establishing LLMs, Neo4j because the default database for improved AI/ML mannequin efficiency, data graphs to boost GenAI predictions, LangChain orchestration for context-aware reasoning functions, and numerous supporting instruments and assets. This initiative goals to empower builders to leverage AI/ML capabilities of their functions effectively and securely.
“Builders are excited by the probabilities of GenAI, however the price of change, variety of distributors, and large variation in know-how stacks makes it difficult to know the place and the way to begin,” mentioned Scott Johnston, CEO of Docker CEO Scott Johnston. “At the moment’s announcement eliminates this dilemma by enabling builders to get began rapidly and safely utilizing the Docker instruments, content material, and providers they already know and love along with accomplice applied sciences on the slicing fringe of GenAI app growth.”
Builders are supplied with straightforward setup choices that provide numerous capabilities, together with easy information loading and vector index creation. This permits builders to import information, create vector indices, add questions and solutions, and retailer them throughout the vector index.
This setup allows enhanced querying, end result enrichment, and the creation of versatile data graphs. Builders can generate numerous responses in several codecs, similar to bulleted lists, chain of thought, GitHub points, PDFs, poems, and extra. Moreover, builders can examine outcomes achieved between completely different configurations, together with LLMs on their very own, LLMs with vectors, and LLMs with vector and data graph integration.