Imagine an AI chatbot that can only remember the last 2 messages in a conversation. What is the most likely effect on the chatbot's responses?
Think about how limited memory affects understanding of ongoing topics.
Short-term memory limits the chatbot to recent messages, so it can lose track of earlier context, leading to less coherent replies.
Given this Python code simulating conversation memory updates, what is the final memory content?
memory = [] # User says hello memory.append('User: Hello') # Bot replies memory.append('Bot: Hi! How can I help?') # User asks a question memory.append('User: What is AI?') # Limit memory to last 2 messages memory = memory[-2:] print(memory)
Remember the memory is limited to the last 2 messages.
The code keeps only the last two messages after appending the third, so the first message is dropped.
You want to build a chatbot that remembers long conversations (hundreds of messages). Which model architecture is best suited for this?
Think about models designed to handle sequences and remember past inputs.
RNNs with gated units like LSTM or GRU are designed to remember information over long sequences, making them suitable for long conversation memory.
In Transformer-based chatbots, which hyperparameter mainly controls the length of conversation history the model can attend to?
Consider what limits the input size the model can process at once.
The maximum sequence length or context window size determines how many tokens from past conversation the Transformer can attend to at once.
You want to evaluate if a chatbot correctly uses past conversation context in its replies. Which metric is most appropriate?
Think about a metric that checks if replies fit the conversation flow.
Contextual coherence measures how well the chatbot's replies relate to previous messages, indicating effective memory use.