What should be a good approach to achieve the context based discussion keeping token limit in mind? #344
                  
                    
                      naveengujjar29
                    
                  
                
                  started this conversation in
                General
              
            Replies: 1 comment
-
| There are two ways I have found. 1st approach is to pass the last "K" conversations with the context window size in mind. 2nd approach is to make use of summarising the previous discussion to compress the information being passed for tracking the historical context of the discussion. | 
Beta Was this translation helpful? Give feedback.
                  
                    0 replies
                  
                
            
  
    Sign up for free
    to join this conversation on GitHub.
    Already have an account?
    Sign in to comment
  
        
    
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I have used the spring-ai for RAG on custom data and use the PineCone VectorStore DB.
With my current implementation, I perform the VectorStore similarity search for the provided input by user then pass that context as well in UserMessage but it is causing issue in "Token Limit Exceed" after some time.
What should be the general practice for such a scenario? How are we going to maintain the chat context?
Below code for reference I have used currently.
Beta Was this translation helpful? Give feedback.
All reactions