TiRex - Time Series mit State Tracking für die Robotik (embedded)

Shownotes

Wir haben einen neuen Podcast Partner: Unser Dank geht an die Hannover Messe

Fragen oder Ideen zur Robotik in der Industrie? helmut@robotikpodcast.de oder robert@robotikpodcast.de

Transkript anzeigen

00:00:00: Robotic in der Industrie, der Podcast mit Helmut Schmidt und Robert Weber.

00:00:09: Hallo liebe Zuhörer und Zuhörer, willkommen zu einer neuen Folge unseres Podcastes "Robotic

00:00:16: in der Industrie".

00:00:17: Mein Name ist Robert Weber und heute gibt es eine besondere Folge Hinweis.

00:00:23: Zuerst, ich bin ein bisschen Bayer, das wisst ihr alle, aber es gibt heute eine Folge mit

00:00:27: dem Professor Dr.

00:00:28: Sepp Hochreiter und der hat nämlich letzte Woche sein neues Zeitrein-Foundation-Modell

00:00:35: vorgestellt mit dem Namen Tyrex und hat damit sofort alle internationalen Leaderboards

00:00:41: erklaommen und ist dort auf Nummer eins vor den großen Alibabas, Amazons, Googles etc.

00:00:47: und dieses neue Zeitrein-Modell könnte auch für die Robotik sehr interessant sein, denn

00:00:54: er kann das sogenannte State Tracking.

00:00:57: Was das alles heißt, wie er das macht, welche Anwendungsfälle er da in der Robotik sieht,

00:01:03: hört ihr jetzt im Podcast mit meinem lieben Kollegen Peter, der die Folge aufgezeichnet.

00:01:09: Also viel Spaß und bis zum nächsten Mal.

00:01:11: Hi there, my name is Peter Sieberg and I am your host.

00:01:16: Today I am going to be talking to the one and only Zapp Hochreiter and Zapp and I are going to be talking about Tyrex

00:01:24: and Tyrex is the first XLSTN based time series Foundation Model.

00:01:31: Hi Zapp.

00:01:32: Hi.

00:01:33: How are you doing?

00:01:34: I am fine, I am very excited about because we are launching Tyrex, I am very excited.

00:01:40: Oh, that's great.

00:01:41: We are going to be talking about Tyrex in just a minute.

00:01:44: We have had you on our show actually two, three times.

00:01:50: And for other reasons I believe that at least 95% of our listeners will have heard about

00:01:56: you, but still maybe very quickly introduce yourself to our listeners.

00:02:01: Yes, my name is Zapp Hochreiter.

00:02:05: I am heading the Institute for Machine Learning here in Linz.

00:02:09: It's a JKU, Johannes Kepler University.

00:02:12: I am also Chief Scientist of newly funded AI company called NXAI and this company is

00:02:23: dedicated to bring AI to industrial applications, to bring AI to the machinery and to be a focus

00:02:32: in its moment at XLSTM, the new technique.

00:02:37: And I am known for inventing LSTM, LSTM stands for Long Short-term Memory and LSTM started

00:02:44: all this chatbot, chatGPT stuff because the first large language model was an LSTM model

00:02:50: and I am known for LSTM.

00:02:52: Great.

00:02:53: Thank you very much.

00:02:54: Last time that we met was actually in Linz.

00:02:56: You referred to Bose, your company, new company NXAI, which you are a co-founder of, as well

00:03:04: as where the Johannes Kepler University is.

00:03:07: Yeah, you already referred to LSTM.

00:03:11: I dare to use the quote "great thinkers stand on the shoulders of giants" and even if it

00:03:19: was themselves.

00:03:21: So why maybe don't you quickly take us by the hand, look back at your, I don't know,

00:03:28: maybe 30, 35 years of AI research.

00:03:33: Maybe you want to tell us what were the main stations that brought you then to XLSTM.

00:03:38: Yes, I invented LSTM 1991 in my diploma thesis.

00:03:45: I first analyzed the vanishing gradient, which is a common problem in deep learning, which

00:03:52: you have to overcome to build large models.

00:03:55: And I proposed LSTM architecture for the current neural networks, for neural networks,

00:04:00: which can process time series, which can process text.

00:04:05: But then neural networks were not popular anymore.

00:04:08: In the community support vector machines came, even we had problems to publish LSTMs.

00:04:16: And then starting in 2006, deep learning came, starting in 2010.

00:04:22: LSTM became very popular, all Text and Speech programs on cellphones, we are LSTM based.

00:04:33: It has many, many LSTM applications, the same with Amazon and you name it Microsoft or

00:04:40: so.

00:04:41: But then it turned out 2017, there's another technique, it's called transformer, where

00:04:48: the tension mechanism is built in, that these architectures are better in parallelizing.

00:04:56: You can push more data through these models in training than you could in LSTMs.

00:05:05: If you have the first LSTM, the first large language models were based on LSTM, but this

00:05:11: parallelization to get more training data at the same time pushed LSTM from the market

00:05:17: and transformer was used.

00:05:20: And I always thought, hmm, can we not scale up LSTM like transformers?

00:05:27: Can we not do the same?

00:05:29: Can we not build large models?

00:05:30: Can we not make it faster?

00:05:33: And with XLSTM, we achieved this.

00:05:35: We looked into it, we copied some of the tricks of the transformer technology, added some

00:05:43: of our own tricks from the LSTM technique and then published this LSTM technology, which

00:05:52: is a model, which is based on original LSTM, but can be paralyzed and has some other tricks,

00:06:03: which make it really, really powerful.

00:06:05: And we showed it can achieve the performance of transformers in large language modeling.

00:06:12: We will show soon that we are the same level as transformers.

00:06:16: Right.

00:06:17: We may be coming into a little bit more detail later on in this comparison of the transform

00:06:23: and XLSTM technology.

00:06:25: But our topic today is time series.

00:06:27: Now, am I correct in assuming that until recently with regards to XLSTM, as you just introduced,

00:06:36: you have been concentrating on language.

00:06:39: So, how is time series data different from non-time series data?

00:06:45: So the data that does not have any timestamp from the perspective of the researcher Sepolreter.

00:06:52: Yes, yes.

00:06:53: Yes, first of all, for me, there's not a big difference.

00:06:57: If you give me a sequence, I can use every time series method because I can assign to

00:07:04: the sequence element time points and I can analyze sequences like DNA or even text.

00:07:11: Also might be a sequence from assist.

00:07:13: There's not a big difference, but the data is different.

00:07:18: If you look at text, there are coordination between words, which are far away and this

00:07:25: is a more abstract symbols, your process.

00:07:30: In timeseries in most cases, you have numerical values, you have numbers or vector and you

00:07:38: process numerical values.

00:07:40: And often in timeseries, the data comes out of a complex system.

00:07:45: The system has something like a hidden state.

00:07:48: It's about in what state is the system and then you want to predict the future or you

00:07:55: want to classify what's happening right now.

00:07:58: This is a difference between abstract symbols, which have some meaning and numerical values,

00:08:06: which came out of a complex system with hidden states.

00:08:09: Right, so referring to the systems, maybe you can give us a couple of examples.

00:08:15: Timeseries are being used in a variety of very different markets.

00:08:20: Maybe you can give us a couple of examples of use cases and markets where the typical

00:08:25: timeseries data comes from.

00:08:29: Timeseries pervasive, so everywhere you find them everywhere and you encounter them everywhere.

00:08:36: If you think about weather forecasting, if you drive with your car and the navigator

00:08:43: tells you as estimated to time of arrival, it's a timesherring, it's forecasting.

00:08:48: If your system tells you when the battery, if you have an e car, is empty, it's a time

00:08:54: series problem, but it's in stock market prediction, in predictive maintenance, in

00:09:01: logistic you have to predict.

00:09:03: When do you have to order new parts, set your production, that's still.

00:09:09: Or when your machinery needs new oil, you have to predict some market.

00:09:16: For example, if you produce something for the car industry, you have to predict how many

00:09:22: cars will be sold to a trustier production.

00:09:27: Very prominent was Amazon.

00:09:31: They have all across timeseries prediction because they have to predict two things first,

00:09:39: how much a product is bought and also how long does it take to deliver it.

00:09:45: Because they have some deliveral things, say a better in predicting how well a product

00:09:50: is sold, so producers themselves.

00:09:54: Amazon is one prediction company and the whole business model is built on a prediction.

00:10:00: But you need it for climate, you need it for medicine, there's EEG and EKG, there are so

00:10:08: many predictions.

00:10:10: You want to know how the party is responding to treatments or during a surgery.

00:10:18: As applications in agrar, if you do some corn or apples or whatever, you have to predict

00:10:27: the weather, you have to predict the soil condition.

00:10:31: A very famous application, where we are very good, is hydrology to predict floodings.

00:10:40: Because here, if it's raining, we have these hidden states, the rain goes into the soil,

00:10:47: goes underground, basins, underground basins, and you have to memorize how full are these

00:10:54: basins, because if they are full, the rain directly will go into the river.

00:10:59: Assobuy the underground basins will be filled up before the water goes into the rivers.

00:11:04: This is a very, very prominent example of how we do earth science in climate change,

00:11:11: where you need this forecasting all the time.

00:11:13: You need forecasting in energy, smart grids, you have to predict the weather for solar

00:11:19: energy, wind energy, and you have also to predict the customer behavior.

00:11:24: If there's something like a football game, like Germany is in the final, everybody turns

00:11:31: on the TV and puts a beer into the fridge or whatever.

00:11:35: This is a couple of examples, but there are so many, many, many more.

00:11:39: It's everywhere, it's really everywhere.

00:11:41: Ja, really.

00:11:42: We hear you and I'm sure you could go on for a couple of minutes.

00:11:45: Yeah, so very good.

00:11:46: You gave a couple of examples of the specific area.

00:11:50: We have a main interest in here in our podcast in the industrial environment.

00:11:56: Since when then, looking back these, whatever, 35 years, since when have you been looking specifically

00:12:03: at time series?

00:12:04: This is from the moment that you came up with LSTM.

00:12:08: If that was the case, what were until then the main algorithmic capabilities will come

00:12:15: to the new ones later on, but what were the standards in the past that were capable of

00:12:20: looking into the future of time series?

00:12:23: Yes, I started in Kindergarten.

00:12:26: I was always interested to predict the future, but now kidding, this LSTM, the first LSTM

00:12:32: applications were time series because text was not available.

00:12:36: We never thought about doing text LSTM and where I come from.

00:12:42: Only time series were in our mind.

00:12:45: An LSTM has been designed for time series, so old original LSTM and it performed very

00:12:52: well.

00:12:53: LSTM is used everywhere.

00:12:56: Even one guy from Google told me LSTM is still used in Google Translate because its faster

00:13:03: than the transformer architecture in InfoRence, in applying it.

00:13:07: But LSTM were in many, many industries, in many, many broad domains, in industry for prediction.

00:13:16: I gave a couple of applications, but there are many more.

00:13:21: LSTM was good there.

00:13:23: Alternatively, there were models like RIMA, statistical models, so I only do this local

00:13:30: averaging.

00:13:31: Meaning, you make an average over the last values or you calculate a trend or something

00:13:39: like this.

00:13:40: This was typically for stock market predictions with traditional statistical methods.

00:13:45: And LSTM was better because LSTM could memorize stuff and it could memorize in what state some

00:13:53: system is.

00:13:54: I brought you a hydrology thing here.

00:13:57: If it's snowing, the snow does not go to water.

00:14:00: So the snow is stored, the snow lying on the soil.

00:14:04: And if the sun shines, the snow tries into water.

00:14:07: And this is something like storing water, also in the Gletscher underground presence.

00:14:12: Some systems also, the sea, if there's a storm, it's a sea, you don't see it.

00:14:18: But there's a hidden state because in the sea, under the water, still a lot of food is in

00:14:24: the water because of the storm before.

00:14:27: In fish eating this, there are these hidden states everywhere.

00:14:31: And these statistical methods were not good to capture the hidden states because they

00:14:36: do it on your averaging.

00:14:37: LSTM was very good to capture the hidden states of some systems.

00:14:42: Think about a pipe, you have a water pipe, you open something, water is flowing.

00:14:48: But on the other end, it takes a time until the water arrives.

00:14:53: But you have to memorize, yes, I opened the water pipe.

00:14:56: I opened and the water is flowing.

00:14:58: Sister hidden states.

00:15:01: Very good.

00:15:02: Now you have come with a new time series Foundation model called Tirex.

00:15:09: The king of time series, I assume that's what you want to convey with that.

00:15:13: And it's based on XLSM.

00:15:15: You just introduce XLSM in the comparison with a transformer.

00:15:19: But what are the main features?

00:15:21: What is the USB of Tirex?

00:15:25: Tirex, indeed, it's the king of time series.

00:15:29: It's the king of time series models.

00:15:31: First of all, it's based on XLSM.

00:15:34: And I already told you, so original LSTM is very, very good in time series prediction.

00:15:40: Now we improved it.

00:15:42: But it still kept its super performance in time series prediction.

00:15:48: It's very good.

00:15:49: But with all these tricks of the transformers, it became even more powerful.

00:15:56: And this is a time series Foundation model.

00:15:59: What does this mean?

00:16:00: This is a new kind of time series prediction, which come out of this large language models

00:16:08: because of the in context learning.

00:16:10: For large language models, you can write something in the context, give some questions

00:16:16: or give some examples, and then the large language models is processing this and gives you an answer.

00:16:22: Here's the idea.

00:16:23: I train a very large model on many, many different time series.

00:16:28: And then I give a new time series in context.

00:16:31: It's like a prompt.

00:16:32: It's like a question.

00:16:33: But in this case, only numerical values.

00:16:37: It's a type series.

00:16:38: And then you say, can you give me the future?

00:16:41: Can you give me the next time point or the next 10 time points?

00:16:44: Or can you give me what's happening in 100 time points?

00:16:48: And this idea of the large language models has so much knowledge.

00:16:56: And this time series Foundation models have so much knowledge about time series.

00:17:04: Since I don't have to learn new time series, but I already see patterns.

00:17:09: I saw an other time series, and if we give a prefixes, a beginning of a time series,

00:17:15: for them it's clear, yes, as a future will look like this.

00:17:18: Here we have a very, very, hutsing CIS Foundation models.

00:17:22: First of all, so allow non experts to use high quality time series models.

00:17:29: You have no idea about time series.

00:17:31: You put it in context, all your values, and you get good prediction.

00:17:34: Wow, you don't have to know anything about time series or deep learning.

00:17:40: That's the first big advantage.

00:17:42: So second big advantage is, if you don't have enough data,

00:17:47: then you cannot learn a model for your particular domain or time series.

00:17:53: But this Foundation model, you only give the beginning of your time series.

00:17:57: And you don't have any data.

00:17:59: You don't have training data, but the model already makes good prediction.

00:18:03: So, Jeffins, perfectly suited for tasks where not enough data is available.

00:18:11: Okay, very good.

00:18:13: So, what about, so this is like about the quality, maybe the use, we come to that in a moment.

00:18:19: At the very end, we're going to be looking at some benchmark numbers, maybe do some comparison as well.

00:18:26: But before then, if you compare, what about the size of the model?

00:18:31: What about the speed of the model in relation to other solutions in the market?

00:18:35: Okay, go later to numbers, but compare to other solutions.

00:18:39: I have to mention other solutions.

00:18:41: Almost other competitors in this domain, meaning time series Foundation models,

00:18:49: are based on the transformer technology, because it's so popular,

00:18:53: it's so successful in large language models in Jegebelky, Innoit.

00:18:58: And they have a problem, they have a problem, because they are typically very large

00:19:04: and they are typically very slow.

00:19:08: For example, if you give a time series, as I said in context,

00:19:12: they always, for every prediction, they have to go over the whole time series again and again.

00:19:17: So super slow.

00:19:19: What we achieved is two things.

00:19:22: First of all, our model is small.

00:19:25: Our model has, because it's based on X-Listam, a fixed memory.

00:19:31: So far, it's perfectly suited for embedded systems at edge devices,

00:19:36: which transformer cannot do.

00:19:38: And we are super fast.

00:19:41: We are super fast because of two reasons, because we are small.

00:19:44: Okay, if we are small, we are faster, because we don't have to do so much computations.

00:19:49: But because it inference transformers a quadratic in the context length,

00:19:56: in the length of the time series you give in context.

00:19:58: And the L-Listam is linear, because it only accesses the memory.

00:20:03: It's better.

00:20:04: Well, it's faster, it's much faster.

00:20:07: It's small and faster.

00:20:08: And now, the most important thing is,

00:20:12: it's even better in prediction quality, in forecasting quality, because the XLSM we use

00:20:20: is able to do stage tracking. I told you there are states like in hydrology, if you want to predict

00:20:28: how much water is in your river, there are these hidden states, water is in the snow, water is in

00:20:33: the soil, water is in the underground basins, and you have to keep track of this, you have to memorize

00:20:39: it, you have to track out its raining, but the water is going into the soil, but it will flow out

00:20:44: later, and CISAR states, CISAR hidden states of the system, also in the robotics state would be

00:20:52: where's your robot arm, you can memorize what movements you have done and where your robot arm

00:20:58: is located, and LSTM can do that, but transformers or these fast models like RWKV or Mamba, these

00:21:11: models which came out cannot do the stage tracking, cannot keep track or cannot monitor

00:21:18: in which state your system is, and that's so important, and therefore we are in many times here

00:21:25: is so much better because we can do stage tracking, we can memorize in what state a complex model is,

00:21:33: and to come to the competitors, our competitors are something like Kronos from Amazon,

00:21:39: Times FM from Google, Moray from Salesforce, Toto from DataDoc, and also Alibaba,

00:21:53: the Chinese company put some new foundation models for time series only a couple of days ago

00:22:01: into the hacking phase leaderboards, and just the big companies, they devoted a big team to get good

00:22:08: models, and we are considerably better, we are clearly better than all these methods, because

00:22:16: we have an advantage, because we can do the stage tracking, and it's not only a small difference,

00:22:21: it's a clear difference where we are better, and all these big companies could not keep up with us

00:22:28: because it's a technology, it's our technology, it's a NXCI technology, it's European technology,

00:22:35: which has beaten everything else, and we are not only better in forecasting, as already said,

00:22:43: we are faster and we are smaller, and this is fantastic, that's unbelievable, we are better,

00:22:50: faster, smaller, and we are so happy, we are so excited, that we are clearly in front compared

00:22:58: to these teams of these big companies. That's great, we can really feel your excitement,

00:23:05: Zepp, that is really great, higher quality, more speed, smaller, what does that mean, you

00:23:11: already refer to edge as a potential, maybe give us a couple of typical use cases where you see

00:23:20: Tirex to be applied. Tirex should become a standard, if you do some time series workouts,

00:23:28: then add some machinery, if you have a small device, and you want to know what's happening on

00:23:35: your machine, you do the better control stuff, you should use this, because add some machinery,

00:23:41: you have to be fast to interfere fast enough, and you have to be small, because you cannot

00:23:47: put a big computer besides your machinery, small and fast is important, and being good

00:23:53: is also an advantage, or in process control, like a digital twin, you have a simulation,

00:23:59: and you do prognosis, you do forecasting of your system, like the heat, is it too hot at some point,

00:24:09: if it's too hot, if the forecasting said it will become too hot, you have industrial process,

00:24:14: you have this small device on the side with Tirex in it, Tirex says, hey, stop, it's becoming too

00:24:22: hot, send you a regular down, or Tirex tells you as a catalystator is not well distributed, because

00:24:30: the forecasting can predict the distribution of the catalyst or some chemical material in your

00:24:37: process, says, hey, we have to change this, give more of it, or whatever, and this is important,

00:24:44: because this has to be in real time, if you want to steer the process, if you want to control the

00:24:51: process, it has to have real time capacities, it has to be small, because they have to fit into a

00:24:57: small device, an embedded device in your production system, but also Tirex, you will see it in

00:25:04: autonomous driving, because in cars, you have to predict when is the battery empty, and there are

00:25:10: many prediction things, you will see it in drones, if you have to predict it, you will see it in all

00:25:17: autonomous systems, especially in autonomous production systems, because Tirex is good, Tirex,

00:25:26: I mean, it's a quality of prediction, is small, it fits on small devices, and it's super fast,

00:25:32: yes, that's ideal for industry, industry should jump on it. Exactly, and I'm so happy, and I'm

00:25:40: sure that many listeners are so happy hearing exactly this, it's almost like as if you have

00:25:47: produced, you know, we started working three, four years ago, and now you come with this

00:25:52: great solution, almost as if it was specifically made for our audience, so to say, very good, so

00:26:00: you already referred to, it will be telling you, who is the you, I mean, you refer to the continued

00:26:07: state tracking, but also about the context learning specifically, so what does that mean,

00:26:13: who is going to be the typical user, is that changing, is it more the data scientist type of

00:26:19: very knowledgeable person, or does it mean that you're going to have like typically the domain

00:26:24: expert being capable of using solutions that are going to be based on Tirex? That's a good thing,

00:26:31: because you don't have to be an expert anymore, because you download your Tirex, you feed your

00:26:38: numbers, your time series into the context, and you get a prediction, and the prediction

00:26:43: is as good, and in most cases even better, than if you would build a model, use also expert knowledge

00:26:53: and time series research, and do a prediction, that's super good, because now time series prediction

00:27:00: is open for everybody, but even better, even better, assume you are a company, and you sell

00:27:08: a device to different customers, every customer says, can you adjust the device to my needs,

00:27:15: can you adjust the device to my environment, or to my product, or whatever, and then you

00:27:21: need somebody who is mind-tuning the time series prediction model, or as a forecasting model for

00:27:27: each customer. If you use Tirex for example, you put Tirex on it, on the machinery, it goes to the

00:27:35: customer, as the customer starts the machinery, Tirex will suck in the data from the machinery,

00:27:45: and put it in context, and it's doing prediction, and if the customer has a new product, Tirex will

00:27:51: get in the data for the new product, or the new use of the machinery, and can do prediction,

00:27:59: if the machinery is worn out or changes its behavior, Tirex can get in the actual data,

00:28:06: and do prediction, and you sell something, and you don't have to care about it, because Tirex

00:28:14: can adapt to all changes, because it can automatically load time series into the context,

00:28:23: and track the machinery, track the use of the machinery, and you don't have to do anything

00:28:28: anymore, as a company is selling machinery with time series forecasting in, built in,

00:28:34: in the machines you're selling. That's really great here, it's a direction that I've been

00:28:40: looking at and expecting, almost like for quite some time, that domain experts are going to be

00:28:47: using their data, the data that they've been producing, but were never capable of doing

00:28:54: something with themselves, always needed to go to other people, third parties or in-house.

00:29:01: Now, you gave a general example of a company selling devices, now what is going to be the type

00:29:09: of Tirex customer, what kind of product or service are they going to build on top of Tirex,

00:29:18: or are they going to be using Tirex directly, and maybe you want to tell us then in relation to

00:29:23: that, what is going to be the type of license that you're going to put Tirex onto the market.

00:29:28: First of all, Tirex is a base model, we will put on Hackingface to show everybody

00:29:36: that we have better, better sensors, Amazon guys, Google guys, Salesforce guys, Data guys,

00:29:43: Alibaba guys, you name it, a better sensor, American sensor, Chinese, so we have to go out,

00:29:48: but what we can do then is to fine-tune, so base model can do every time series,

00:29:56: but if you have enough data in one domain, you can a little bit tweak and you always get better

00:30:02: in this specific domain if you are trusted, and there are tricks, how to do fine-tuning,

00:30:09: how to trust it to a specific application, so you get better. So, basic model is already better

00:30:17: than specific models used by statistic guys, what is used right now, but you can get even better

00:30:26: if you do fine-tuning, fine-attachment if you go into your domain, and this would be customers,

00:30:34: you may say we have the space model, but we can adapt it to your use case and you get even better

00:30:41: performance, perhaps you get even faster, you can address it to your hardware, to your chip,

00:30:48: to your embedded device, and here as a customer will pay us, hopefully said we adapt this super

00:30:59: cool model, it's super strong to their hardware, their specific applications. Very good, talking

00:31:07: about the specific data, I understand, so there's going to be, I don't know, there's going to be a

00:31:13: hydraulic model, there's going to be a whatever type of machine robotic model, etc. Now the model

00:31:21: that you come with, which is already very powerful, that was based on available public data or maybe

00:31:29: also on data from companies that you have been working with in specific industrial segments?

00:31:35: Right now it's only based on public data, it's important because otherwise we would have a

00:31:42: license problems, it's based on public data, and here a nice thing is a couple of days,

00:31:48: a new model came out, it's called Toto from DataDoc, a big American company, and they had one

00:31:58: trillion internal data, additional to the public data we are using, and we're still better, that's

00:32:07: like a joke, because they used internal data to build their model, additional to the data we

00:32:14: have available, imagine if we would have all the data the companies have internally, we're beating

00:32:22: them, but what model we would build if we would have also access to this data, it would be unbelievable,

00:32:30: and here we hope that we get more data from our industrial partners to even build on top of this

00:32:39: Tyrex model, even better, more specific models, like multivariate data, we already have ideas how

00:32:47: to make it multivariate, stuff like this, but here for buildings we need good data, and we are

00:32:55: right now collecting data, we are right now asking different partners, can we collaborate to build even

00:33:03: stronger time series models, but we are so strong already, but we are looking into the future.

00:33:08: You can become even better, and I'm sure there's going to be hundreds, if not thousands of listeners

00:33:14: of companies that are going to be very, very much interested in using one way or the other

00:33:20: their data in combination with your Tyrex. Okay, let's look a little bit at the numbers, you refer to

00:33:27: two or three competitors, let's say in the market already, maybe you want to share with us what are,

00:33:34: what is the number one, or maybe number two, three time series benchmarks, and you refer to two,

00:33:41: three potential competitors, and maybe you want to tell us then how Tyrex is performing relative to them.

00:33:50: Yes, it's a little bit complicated because there are some evaluation measures, and if you're not

00:33:56: familiar, there are only numbers for you. Let's say if we go back to the status we were seeing,

00:34:04: now there are new submissions, there's one measurement method, it's called CRPS, it's about

00:34:12: probabilistic forecasting, where you not only say one point, but you say some, like an interval,

00:34:19: you know how good it is, and there was with these numbers, the smaller numbers are better,

00:34:26: the CRONOS had 0.48, CRONOS is from Amazon, times FM from Google has 0.46, TAP, PFN, that's a

00:34:39: method from Franco to in Freiburg has 4.8, all these methods, foundation methods, there's also

00:34:48: Morai from Salesforce, Salesforce invests a lot into time series, it was about 0.5, and you see

00:34:59: all lined up at 4.6, 4.7, 4.6, 4.7, and we get into the same measurement, 4.1.

00:35:08: There's a big gap, it's all big companies are competing at the level of 4.6, 4.7,

00:35:17: and with our first submission, we got 4.1, you see it's a gap, another method, another criteria

00:35:26: would be the rank, you don't do the evaluation on one time series, you go over many, many time

00:35:33: series, and then you want to see how good you are on what rank you are, what is the average rank,

00:35:40: perhaps you're one second, then you're third, then you work first, and you give the average rank,

00:35:46: and if you do this average rank on what place you are, we get for Tyrex, we got on average 3

00:35:55: over many, many, many methods, also specialized methods, and the next best method like time

00:36:03: is 6 on average, is on place 6 on average, Kronos is on place 7 on average, Morai is also on place

00:36:11: 7 on average, we are on 3, and the next ones are on 6, you see there's a big gap, whether you

00:36:19: measure directly the performance, the prediction performance, while you rank the method say on

00:36:25: what place you are, are you first or second, and then average over the places, we are also with

00:36:31: a big gap better than all others, it's so fantastic, we couldn't believe it, that we are performing so,

00:36:38: so good, and the reason for this is the technologies that you refer to, continued

00:36:45: state tracking, context learning in combination with on top of XLSDM in relation to all the other

00:36:52: ones are transformer based or is that not necessary, all others now transformer based,

00:36:58: because transformers are so popular, but in industry, in practice, LSDM performed very well,

00:37:05: LSDM was always strong in time series, but more than transformer based methods, but this is now

00:37:12: in context learning, said you don't learn, which is known for large language models,

00:37:18: and therefore everybody jumped onto transformers, because we know transformers can do this,

00:37:23: if it was not clear, can LSDM or XLSDM do this, said XLSDM can do it, it was for me, it was clear

00:37:31: because we are also in language, but here it went through the roof with this performance.

00:37:37: Very good, congratulations, it sounds really, really impressive, before we're going to close

00:37:44: off, why don't you share with us maybe where your base, where's your team, both for NXAI as well as

00:37:52: your job at the Johannes Kepler University, maybe you're looking for new colleagues,

00:37:58: there's jobs open maybe, and if so what should interested people bring?

00:38:02: Yes indeed we have jobs open, we are located in Linz, both as a company NXAI is in Linz,

00:38:11: also my institute at the university is in Linz, we are looking always for very motivated,

00:38:19: interested researchers, but also developers, it's such an exciting field, believe me, if you join

00:38:27: us you will have fun, it's just great to do it and also many success stuff, what we also offer is

00:38:35: dual systems that you can also work from home half of the time or something like this,

00:38:41: this would be a negotiated and we have a very inspiring environment, many researchers, many

00:38:49: new ideas, everything is on fire. That's amazing, maybe I'll consider applying for a job with you,

00:38:57: no I will not, but you dear listener, I'm sure that there's going to be many many people

00:39:04: and I think the most important thing is you kind of, but also you are too modest, but we can all

00:39:11: feel your excitement and we here, we heard again today with the great technology coming from you,

00:39:19: coming from Linz from also coming from Europe, so and I can only support you and suggest any

00:39:27: interested party person listening to be contacting you, so Zapp thank you very very much again,

00:39:35: as I suggested before it feels almost like you are now so close to our industry, to our industrial

00:39:43: environment here, we are very very much looking forward to seeing solutions based on Tirex,

00:39:50: the time series foundation model that is better, smaller and faster, thank you very much Zapp and

00:39:57: looking forward to see you soon in the Alps again. Yes, it was a pleasure and please check out Tirex,

00:40:05: it's rewarding. Thank you Zapp, bye bye. Bye bye, ciao.

00:40:20: (gentle music)

00:40:22: [BLANK_AUDIO]

Neuer Kommentar

Dein Name oder Pseudonym (wird öffentlich angezeigt)
Mindestens 10 Zeichen
Durch das Abschicken des Formulars stimmst du zu, dass der Wert unter "Name oder Pseudonym" gespeichert wird und öffentlich angezeigt werden kann. Wir speichern keine IP-Adressen oder andere personenbezogene Daten. Die Nutzung deines echten Namens ist freiwillig.