[1] Anthropic Research Team. (2025).
Tracing the thoughts of a large language model. Anthropic Research Blog.
https://www.anthropic.com/research/tracing-thoughts-language-model[2] Anthropic Research Team. (2025).
Circuit Tracing: Revealing Computational Graphs in Language Models. Transformer Circuits Publication.
https://transformer-circuits.pub/2025/attribution-graphs/methods.html[3] Anthropic Research Team. (2025).
On the Biology of a Large Language Model. Transformer Circuits Publication.
https://transformer-circuits.pub/2025/attribution-graphs/biology.html[4] Cunningham, H., Ewart, A., Riggs, L., Huben, R., & Sharkey, L. (2023).
Sparse Autoencoders Find Highly Interpretable Features in Language Models. arXiv:2309.08600v3.
[5] An, Y., Yang, B., Zhang, B., Hui, B., Zheng, B., Yu, B., Li, C., Liu, D., Huang, F., Wei, H., Lin, H., Yang, J., Tu, J., Zhang, J., Yang, J., Zhou, J., Lin, J., Dang, K., Lu, K., Bao, K., Yang, K., Yu, L., Li, M., Xue, M., Zhang, P., Zhu, Q., Men, R., Lin, R., Li, T., Tang, T., Xia, T., Ren, X., Fan, Y., Su, Y., Zhang, Y., Wan, Y., Liu, Y., Cui, Z., Zhang, Z., & Qiu, Z. (2025).
Qwen2.5 Technical Report. arXiv:2412.15115v2.
[6] Es, S., James, J., Espinosa-Anke, L., & Schockaert, S. (2025).
Ragas: Automated Evaluation of Retrieval Augmented Generation. arXiv:2309.15217v2.