Bitcoin neural network analysis week 33/2019: The Verdict

in #bitcoin5 years ago

Analysis of past week:
Last week was a bit of up and down. The network strategy was suboptimal.

I am comparing three strategies, first hold fiat and make a small profit elsewhere, hold bitcoin, or follow the network predictions. Since I started documenting my predictions on the blockchain these strategies compare as

Fiat with a weekly profit of 0.5%: 16727.17$ (up 0.5 %)
Holding BTC: 11523.58$ (up 4.9 %)
Network strategy: 11273.71$ (up 0.5 %)
Starting capital: 11497.40$

I have been doing this for more than a year now and I think it is time to draw some conclusions.
The bitcoin price is now pretty much where I started. In-between we had a fair amount of ups and downs. And the neural network is exactly on par with hodling bitcoin. There was nothing gained by using the network and (almost) nothing lost.
And of course since we are not in a raging bull market, fiat with 30% profit is superior.

I am not convinced that this is just bad luck on part of the network, but that there is a more fundamental problem. Neural networks find patterns. But are these patterns just statistical flukes that will never be repeated in the future, or are they actual based on real dynamics. A network (at least the type I am using) can never distinguish those two. In addition the training data is extremely limited and it is not simple to build much more sophisticated networks.

Therefore I am no longer convinced that using neural networks is the way to go. Instead I am now pursuing more statistical methods. I will post more when I am finished, but since I no longer believe in the network I will stop posting these forecasts.

Sort:  

Posted using Partiko Android

I've reproduced 130+ research papers about "predicting bitcoin", coded them from scratch and recorded the results. Here's what I've learnt. ok, so firstly, all of the papers I found through Google search and Google scholar. Google scholar doesn't actually have every research paper so you need to use both together to find them all. They were all found by using phrases like "predict bitcoin" or "predict stock market" or "predict forex" and terms related to those. ​ Next, I only tested papers written in the past 8 years or so, I think anything older is just going to be heavily Alpha-mined so we can probably just ignore those ones altogether. ​ Then, Anything where it's slightly ambiguous with methodology, I tried every possible permutation to try and capture what the authors may have meant. For example, one paper adds engineered features to the price then says "then we ran the data through our model" - it's not clear if it means the original data or the engineered data, so I tried both ways. This happens more than you'd think! ​ THEN, Anything that didn't work, I tried my own ideas with the data they were using or substituted one of their models with others that I knew of. ​ Now before we go any further, I should caveat that I was a profitable trader at multiple Tier-1 US banks so I can say with confidence that I made a decent attempt of building whatever the author was trying to get at. ​ Oh, and one more thing. All of this work took about 7 months in total. ​ Right, let's jump in. ​ So with the papers, I found as many as I could, then I read through them and put them in categories and then tested each category at a time because a lot of papers were kinda saying the same things. ​ Here are the categories: * News Text Mining. - This is where they'd use NLP on headlines or the body of news as a signal. * Social data - Twitter Sentiment/Google Search/Seeking Alpha. Again, some were NLP, for google trends they just used the data. * Technical Analysis & Machine Learning together. Most of these would take the price, add TA features, then feed into a ML model. * Other machine learning (as in, not using TA). Just using the price and some other engineered features. * Analyst Recommendations. Literally just taking the recommendations from banks/brokers and using that as the signal. * Fundamental data. So ratios from the income statement/balance sheet, ​ Results: Literally every single paper was either p-hacked, overfit, or a subsample of favourable data was selected (I guess ultimately they're all the same thing but still) OR a few may have had a smidge of Alpha but as soon as you add transaction costs it all disappears. ​ Every author that's been publicly challenged about the results of their paper says it's stopped working due to "Alpha decay" because they made their methodology public. The easiest way to test whether it was truly Alpha decay or just overfitting by the authors is just to reproduce the paper then go further back in time instead of further forwards. For the papers that I could reproduce, all of them failed regardless of whether you go back or forwards. :) ​ Now, results from the two most popular categories were: * *Social data.*A lot of research papers were extensions of or based off of a paper by Johan Bollen called "Twitter mood predicts the stock market". It literally has 3,955 citations and is complete and utter horse shit; the paper is p-hacking to the extreme. Not only could I not reproduce the results, but given the number of sentiment indicators he uses I regularly found correlations between sentiment and my data based on how I engineered it. None of these correlations held over longer time periods. Every paper that's a derivative of this one or cites it has the same issues. * *Technical analysis & machine learning.*Every paper would do something along the lines of.. take past price data for some asset (stocks, forex), then add technical analysis indicators as "features". Then either they'd run through a feature-selector that figures out the best features then put the best ones into a model OR they'd dump this data straight into the model and afterwards select the subset of instruments that it "worked" on. None of these would hold if you k-fold test them or test on different subsets of data outside of the ones used in the paper. The results are always based off of selecting favourable subsets of data. ​ The most frustrating paper: I have true hate for the authors of this paper: "A deep learning framework for financial time series using stacked autoencoders and long-short term memory". Probably the most complex AND vague in terms of methodology and after weeks trying to reproduce their results (and failing) I figured out that they were leaking future data into their training set (this also happens more than you'd think). ​ The two positive take-aways that I did find from all of this research are: 1. Almost every instrument is mean-reverting on short timelines and trending on longer timelines. This has held true across most of the data that I tested. Putting this information into a strategy would be rather easy and straightforward (although you have no guarantee that it'll continue to work in future). 2. When we were in the depths of the great recession, almost every signal was bearish (seeking alpha contributors, news, google trends). If this holds in the next recession, just using this data alone would give you a strategy that vastly outperforms the index across long time periods. ​ Hopefully if anyone is getting into this space this will save you an absolute tonne of time and effort. ​ So in conclusion, if you're building trading strategies, simple is good :) ​ Also one other thing I'd like to add, even the Godfather of value investing, the late Benjamin Graham (Warren Buffet's mentor) used to test his strategies (even though he'd be trading manually) so literally every investor needs to backtest regardless of if you're day-trading or long-term investing or building trading algorithms.

Posted using Partiko Android

you did all of that work? or are you simply citing?

Anyways I pretty much agree with the text.

I am just citing, but I find it relevant and interesting nonetheless.

Thanks for the forecasts so far. It was a nice experiment many of us wanted to try.

Posted using Partiko Android

1 !BEER Token for you

The predictions gave a hint for some gambling. That was fun. I'm curious what you'll come up with.

Posted using Partiko Android

"He" ain't coming up with anything. It's an algorithm that makes the predictions.

Posted using Partiko Android

Hi,
I meant the last paragraph.
Greets

Posted using Partiko Android

Hi @frdem3dot0!

Your post was upvoted by @steem-ua, new Steem dApp, using UserAuthority for algorithmic post curation!
Your UA account score is currently 3.741 which ranks you at #5360 across all Steem accounts.
Your rank has dropped 45 places in the last three days (old rank 5315).

In our last Algorithmic Curation Round, consisting of 149 contributions, your post is ranked at #144.

Evaluation of your UA score:
  • You're on the right track, try to gather more followers.
  • You have already convinced some users to vote for your post, keep trying!
  • Try to work on user engagement: the more people that interact with you via the comments, the higher your UA score!

Feel free to join our @steem-ua Discord server