A research assistant agent receives multiple tasks such as summarizing papers, extracting data, and generating reports. How does it decide which task to do first?
Think about how a helpful assistant would manage multiple jobs efficiently.
A research assistant agent prioritizes tasks by considering deadlines and how important each task is to the main research goal. This helps it work efficiently and deliver useful results on time.
What will be the output of the following Python code simulating a research assistant agent's task queue?
tasks = ['summarize paper', 'extract data', 'generate report'] completed = [] while tasks: task = tasks.pop(0) completed.append(f"Done: {task}") print(completed)
Look at how tasks are removed and added to completed.
The code removes tasks from the front of the list and appends a 'Done:' message to completed in the same order, so the output list matches the original task order with 'Done:' prefix.
Which model type is most suitable for a research assistant agent tasked with summarizing scientific papers?
Think about which model handles text and language well.
RNNs and transformer models are designed to process and generate text, making them ideal for summarizing papers. CNNs and clustering are not suited for text summarization tasks.
Which hyperparameter adjustment is most likely to improve the quality of summaries generated by a transformer-based research assistant agent?
Think about how the model understands longer texts.
Increasing the maximum sequence length allows the model to consider more of the input text at once, improving summary quality. Too high learning rate or zero batch size cause training issues, and fewer attention heads reduce model capacity.
A research assistant agent generates summaries of research papers. Which metric best measures how well the summaries capture the original content?
Think about metrics used in language generation tasks.
BLEU score measures how closely the generated summary matches a reference summary, making it suitable for evaluating summarization quality. Accuracy and confusion matrix apply to classification, and mean squared error on embeddings is less interpretable here.
