The team includes at least 3 chess programmers. Matthew Lai, the author of Giraffe and Talkchess member, is one of them. It is maybe for a reason that Giraffe, following the very same approach as Alpha, is rated only around 2400 on single core.
So, what you are saying is that AlphaZero is 3600 because they have someone on their team who has created an engine that reached 2400?
Likewise, the team has no chess player at above amateur level.
However, the reason Matthew Lai is on the team is that he had tried to produce a chess AI, just one that was 1200 points weaker. 1200 points is not a difference that is achievable by speeding up hardware, even a lot. From articles on this, he was using a much smaller neural network, which even on slower hardware was able to look at about 10% as many nodes as a conventional engine. (see this article)
However, I would agree that using modest computational resources would have been a huge barrier to the development of AlphaZero. The most demanding phase is the self-learning, and this would have taken months without the fast hardware, rather than 4 hours.
The reason AlphaZero benefits from more computation when playing is simply that its search tree gets bigger. But this search tree had 1000 times fewer nodes than that of Stockfish with the exact hardware each used!
It is the huge hardware that made the difference and not the approach.
Self-learning, self-learning, what do you mean self-learning and AI.
You admit you know nothing about the techniques that AlphaZero used to generate its strength: model-based reinforcement learning, termed "deep reinforcement learning" because the model used is a deep neural network.
I have studied this subject (including watching David Silver's excellent lecture series), and use Sutton's book on the subject.
It is true that AlphaZero uses a lot of processing power to achieve its highest strength in head to head play. However, with 30 times less power it would remain the highest rated engine according to AlphaZero's testing. While restricting AlphaZero's computational power would make the match closer, increasing the computational resource for both AlphaZero and a conventional engine like Stockfish would greatly advantage AlphaZero.
The key reason AlphaZero increases in strength more rapidly with computational resource appears to be that the branching factor of its search tree is smaller, to an extent which compensates enormously for it looking at far, far fewer positions. With the full power of 4 TPUs, AlphaZero was still looking at 1000 times fewer nodes than Stockfish! If its time was reduced by a factor of 30, it would be looking at 30,000 times fewer, and still be stronger!
As a result, when AlphaZero gets more time, its horizon must expand significantly faster than that of Stockfish. This is why not only is it stronger, it also indicates this technology has a permanent edge.
You made a good point that I should emphasise: the role of those with knowledge of conventional chess engines in the design of the tree search algorithm of AlphaZero, which has some commonality with all chess engines. While I am no expert on chess engines, a key strength of AlphaZero is that it is better at allocating resources to different branches, and that this is achieved by the quality of the neural network's estimates of the probabilty that each move is best, which comes entirely from self-learning.
And that is why I don't like the approach, because it is too simplistic. The primary code is very simple, as it seems. What is complex is the tuning network, but that is just hardware.
Chess is much more complex than that, theoretically and that is why Alpha will not make big progress in the future.