Tech
Google Introduces 200M-Parameter Time-Series Foundation Model with 16k Context Length
Google has released a new time-series foundation model featuring 200 million parameters and a context length of 16,000, aimed at improving data analysis capabilities.
Editorial Staff
1 min read
Google's latest time-series foundation model is designed to enhance analytical accuracy through a substantial increase in parameters, totaling 200 million.
The model's extended context length of 16,000 allows for improved handling of larger datasets, which is critical for effective time-series analysis.
This open-source model is available on GitHub, providing developers with the tools to implement advanced time-series solutions in their projects.