I am the new one. I am an electrical engineer based in Germany and I am working on a topic in the wide field of "BigData", "Industry 4.0" and "Artificial Neural Networks". I am afraid I can only say those catchwords, but nothing more

We have a software that is running on Windows. My plan is to speed this thing up by using parallel computing. The parallela seems to be perfect, because it offers a lot of computational power with a low price.
Now I have several questions and I hope to find answers here.
1) Is it possible to run a windows based cluster with parallella? There is a blog entry where a Beowulf Cluster has been build, but that is running on Linux. We don't want to offer a Linux system because so far everything is running on Windows. And because we're not working in a scientific field, most users cannot handle Linux.
2) What do I have to keep in mind when building parallel applications with parallella? My imagination is as follows: the software is running on a main node. The software has an interface to some sort of cluster driver and tells him to perform the current computation on another node. The cluster driver manages the nodes and the queries from the software. In doing so, it would be relatively easy to run an algorithm, that is designed for parallel computing, on several parallela boards. Is this imagination too simple?
3) Am I right that if I am building a cluster with 15 parallela Boards I end up with 240 Epiphany cores (plus 15 ARM cores) for 1500$. The 240 Epiphany cores would be responsible for the computation. If so, is it better to use a Nvidia graphics card with lots and lots of cores or a parallella cluster?
I thank you in advance for your support.
Regards