Skip to content

Generates AI-driven responses using the Ollama API with the Phi3 mini model from Microsoft. Heavyweight implementation with memory for language generation, featuring remote RAG and Pinecone for persistent memory

License

Notifications You must be signed in to change notification settings

mydroidandi/commbase-genai-slm-ollama-phi3-mini-memory-remote-rag-pinecone

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Python Package using Conda Python Version

commbase-genai-slm-ollama-phi3-mini-memory-remote-rag-pinecone

commbase-genai-slm-ollama-phi3-mini-memory-remote-rag-pinecone

Interacts with the Commbase platform by generating AI-driven responses using the Ollama API with the Phi3 mini model from Microsoft. This is a heavyweight implementation with memory for language generation, featuring remote RAG and Pinecone for persistent memory.

Examples

Detailed information about examples can be found in the corresponding examples directory.

Contributors

Thanks to the following people who have contributed to this project:

Contact

If you want to contact us you can reach us at [email protected].

License

This project uses the following license: License.

About

Generates AI-driven responses using the Ollama API with the Phi3 mini model from Microsoft. Heavyweight implementation with memory for language generation, featuring remote RAG and Pinecone for persistent memory

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages