githubのsamjabrahams/tensorflow-on-raspberry-piの
Question on GPU #15に、samjabrahams さんが
Question on GPU #15に、samjabrahams さんが
引用 That's a pretty fun idea! I can imagine it being a good exercise, and useful if it's implemented correctly. For me personally, it'll have to be a back-burner idea, but I'll leave some links that might be useful moving forward with that idea: Mirror of the official Videocore IV architecture reference guide Unofficial Broadcom Videocore IV documentation and samples The programmers guide from that repo More unofficial samples and documentation Implementation of GEMM matrix multiplication on RPI Assembler/disassembler for the RPi QPU First of blog posts by Pete Warden on deep learning on RPI Second (and lengthier) post by Pete Warden I'll reopen this thread to increase discoverability.
で、TensorFlow XLAをやればいいと。。。
TensorFlow r1.3で導入された XLA Plugin で、HLOで最適化されたものを Py-VideoCore を使って、
GPUで処理できるところはGPUで、できないところはCPUでやるという感じに。。。
GPUで処理できるところはGPUで、できないところはCPUでやるという感じに。。。
どうだろうか?