Customer-obsessed science
Research areas
-
April 8, 20266 min readAmazon’s RuleForge system uses agentic AI to generate production-ready detection rules 336% faster than traditional methods.
-
April 7, 202613 min read
-
March 20, 202615 min read
-
March 19, 202611 min read
-
Featured news
-
2026Modern language models (LMs) still rely on fixed, pre-defined subword tokenizations. Once a tokenizer is trained, the LM can only operate at this fixed level of granularity, which often leads to brittle and counterintuitive behaviors even in otherwise strong reasoning models. We introduce ByteFlow Net, a new hierarchical architecture that removes tokenizers entirely and instead enables models to learn their
-
2026Transformers are widely used across data modalities, and yet the principles dis-tilled from text models often transfer imperfectly to models trained to other modalities. In this paper, we analyze Transformers through the lens of rank structure. Our focus is on the time series setting, where the structural proper-ties of the data differ remarkably from those of text or vision. We show that time-series embeddings
-
2026Minimizing the inference cost and latency of foundation models has become a crucial area of research. Optimization approaches include theoretically lossless methods and others without accuracy guarantees like quantization. In all of these cases it is crucial to ensure that the model quality has not degraded. However, even at temperature zero, model generations are not necessarily robust even to theoretically
-
2026Knowledge distillation has become a crucial technique to transfer the capacities of large language models (LLMs) to smaller, more efficient models for practical deployment. While recent work exploits rich information from intermediate states of the teacher model for more effective knowledge transfer, imperfect knowledge from the teacher can also mislead student learning, restricting the student’s generalization
-
ICLR 2026 Workshop on Time Series in the Age of Large Models2026Foundation models promise zero-shot forecasting across domains, yet their effectiveness for cold-start scenarios with zero-inflated distributions remains underexplored. We study cross-domain demand forecasting, predicting outcomes for items launching in new domains without historical data where a substantial fraction of launches (≈ 30%) yield zero outcomes and overestimation carries asymmetric costs. We
Collaborations
View allWhether you're a faculty member or student, there are number of ways you can engage with Amazon.
View all