GPU paralleling question

  • 21 Replies
  • 8366 Views
*

unreality

  • Starship Trooper
  • *******
  • 443
GPU paralleling question
« on: November 18, 2017, 04:23:52 am »
Is there someone here who’s coded GPUs? Some of these GPUs have over 4000 cores, but yet they’re only about 5 times faster at data mining than a cpu in desktop PCs. The GPU clock speed isn’t that much slower than the cpu. The cpu only has about 8 cores. The GPU should be at least 300 times faster unless it has some major limitations. I would have thought GPUs would be great a data mining.

Maybe data mining isn’t a good example. Maybe it needs to use global memory too often. Or maybe the cores need to communicate too much or whatever. So what if each gpu core only used it's local memory. Would the gpu then be hundreds of times faster than the cpu? That's what I'd like to really know.

*

ranch vermin

  • Not much time left.
  • Terminator
  • *********
  • 947
  • Its nearly time!
Re: GPU paralleling question
« Reply #1 on: November 18, 2017, 06:21:10 am »
GPU FAN BOY -> ur reading some poor sources there,  my gtx980 (2000 corer) clears my quadcore by 500 times.

MORE LENIENT TO CPUS ->  there are ways to make a cpu go better,  for example,  a Cpu can do a box blur as fast as a gpu by snakeying a box and adding and taking from it,  and also building box keys can be done on cpu quite well if you read once write many,   *but* a gpu will just naively chew through it with a double for loop in the threading box,    so they are both pretty cool.

My honest opinion, is if GPUs had more ram id never use system code again, because its too taxing on my mind, and u get your operation up and running quicker with less building hassles.

Thats... er...  after youve finished getting through all the horrid documentation and wrote a billion lines just to set the basic system up.  hmm im contradicting myself.
« Last Edit: November 19, 2017, 06:19:48 am by ranch vermin »

*

Marco

  • Bumblebee
  • **
  • 34
Re: GPU paralleling question
« Reply #2 on: November 18, 2017, 08:51:36 am »
Is there someone here who’s coded GPUs? Some of these GPUs have over 4000 cores, but yet they’re only about 5 times faster at data mining than a cpu in desktop PCs. The GPU clock speed isn’t that much slower than the cpu. The cpu only has about 8 cores. The GPU should be at least 300 times faster unless it has some major limitations. I would have thought GPUs would be great a data mining.

You cannot compare GPUs to CPUs that easily. Their architecture and how they operate is completetly different. Also, you cannot compare CPUs just by their clock speed and number of cores. Like a 12 year old Pentium 4 processor with 4GHz could not beat in the slightest a single core of a Coffe-Lake or Ryzen CPU which runs on 2GHz.

*

unreality

  • Starship Trooper
  • *******
  • 443
Re: GPU paralleling question
« Reply #3 on: November 18, 2017, 03:04:23 pm »
GPU FAN BOY -> ur reading some poor sources there,  my gtx980 (2000 corer) clears my quadcore by 500 times.

MORE LENIENT TO CPUS ->  there are ways to make a cpu go better,  for example,  a gpu can do a box blur as fast as a gpu by snakeying a box and adding and taking from it,  and also building box keys can be done on cpu quite well if you read once write many,   *but* a gpu will just naively chew through it with a double for loop in the threading box,    so they are both pretty cool.

My honest opinion, is if GPUs had more ram id never use system code again, because its too taxing on my mind, and u get your operation up and running quicker with less building hassles.

Thats... er...  after youve finished getting through all the horrid documentation and wrote a billion lines just to set the basic system up.  hmm im contradicting myself.

That's great news. Maybe you're doing it the right way. Below is one source that gives tons of data mining gpu and cpu examples using well known data mining benchmark apps. The fastest gpu score is 16032, while the fastest cpu score is 3500. The gpu isn't even 5 times faster.

My Surface Pro 3 tablet that I use here to surf the internet that has an i5 is about 1/7th it's gpu.

There's a youtube video (haven't found the link yet), but the guy shows gpu code along with how long the gpu takes to clear 1<20 (~ a million) float type size. It doesn't get much simpler than that. The loop size was 1<20 (~ a million). When he used just one gpu core, it took a whopping 463 ms! When he used 256 cores it took 2.7ms. What's great, but what's interesting is that a typical desktop pc should take about 5 to 10 ms to do that! Once again we have that 1/5th figure! Why?? I understand gpus are amazing at graphics, but I'm more interested in AI, pattern recognition, etc. BTW, that youtube guy has a lot of gpu teaching videos. So one would expect the guy to know his stuff. Why would it take one gpu core so long to clear a million floats? Sure, when he used 256 cores, it was about 170 times faster, but that's only about 5 times faster than a typical cpu. What am I missing?

Zillions of gpu & cpu data mining benchmarks:
http://monerobenchmarks.info/

*

unreality

  • Starship Trooper
  • *******
  • 443
Re: GPU paralleling question
« Reply #4 on: November 18, 2017, 03:10:04 pm »
Is there someone here who’s coded GPUs? Some of these GPUs have over 4000 cores, but yet they’re only about 5 times faster at data mining than a cpu in desktop PCs. The GPU clock speed isn’t that much slower than the cpu. The cpu only has about 8 cores. The GPU should be at least 300 times faster unless it has some major limitations. I would have thought GPUs would be great a data mining.

You cannot compare GPUs to CPUs that easily. Their architecture and how they operate is completetly different. Also, you cannot compare CPUs just by their clock speed and number of cores. Like a 12 year old Pentium 4 processor with 4GHz could not beat in the slightest a single core of a Coffe-Lake or Ryzen CPU which runs on 2GHz.
That must be the case. They're awesome at graphics, but I haven't yet seen an example where they're much more than 5 to 10 times faster than a typical desktop cpu in terms of number crunching one would find in AI and data mining.

My AI will do a lot of basic arithmetic and RAM reads/writes. Is there anyway to get a gpu to be at least a 100 times faster than say a $2000 desktop pc?

*

Marco

  • Bumblebee
  • **
  • 34
Re: GPU paralleling question
« Reply #5 on: November 18, 2017, 03:41:39 pm »
GPUs are good at computing the same calculation on very large batches of data (e.g. training a neural net). That's where they currently outperform CPUs easily.

*

unreality

  • Starship Trooper
  • *******
  • 443
Re: GPU paralleling question
« Reply #6 on: November 18, 2017, 04:30:05 pm »
GPUs are good at computing the same calculation on very large batches of data (e.g. training a neural net). That's where they currently outperform CPUs easily.
That makes sense, albeit disappointing for me since I don't do neural nets. GPUs are very efficient at that. Data mining is probably more like data crunching. While looking a lot of cpu and gpu specs I couldn't help notice that cpus are typically 50 watts while gpus are about 300 watts, although there are a lot of 150 W gpus and 25 watt cpus. 300 / 50 is 6. That's roughly the benchmark difference between gpus and cpus.

Uggg! I guess there's no free lunch? So if I want my AI to be 100 times faster I might have to buy massive amounts of mini motherboards with truckloads of ram chips. It seems RAM is the bottleneck here, no? GPU has incredibly high memory bandwidth, but that's because they read/write thousands of bits at once. Unless I'm missing something here, my AI can't take advantage of that. It's into finer data, dealing with 8 to 64 bit data types, e.g. cluster priorities. My cluster priority doesn't need a thousand bit precision. Even 7 bits is enough.

What I need is more on the lines of massive parallel memory. Each core should have it's own RAM such that each core RAM is not tied down to other RAM blocks. Are there any circuit designers here? Maybe FPGA can accomplish this, but how fast. Imagine a large FPGA chip that has 100 simple CPUs, and each CPU has it's own RAM built into the FPGA. There could also be a central CPU and RAM that periodically communicates with the other CPUs.

*

ivan.moony

  • Trusty Member
  • ************
  • Bishop
  • *
  • 1729
    • mind-child
Re: GPU paralleling question
« Reply #7 on: November 18, 2017, 04:35:59 pm »

*

unreality

  • Starship Trooper
  • *******
  • 443
Re: GPU paralleling question
« Reply #8 on: November 18, 2017, 04:54:28 pm »
A quantum computer is when we will definitely get the Singularity, but hopefully it won't require that. The movie Automata comes to mind. It will absolutely positively 100% guarantee happen. It's only a matter of when. At least by my definition of what Singularity means, which is when AI will be smart enough to improve itself, and that improved version will improve itself, and on and on.

*

infurl

  • Administrator
  • ***********
  • Eve
  • *
  • 1372
  • Humans will disappoint you.
    • Home Page
Re: GPU paralleling question
« Reply #9 on: November 18, 2017, 08:23:57 pm »
A quantum computer is when we will definitely get the Singularity, but hopefully it won't require that. The movie Automata comes to mind. It will absolutely positively 100% guarantee happen. It's only a matter of when. At least by my definition of what Singularity means, which is when AI will be smart enough to improve itself, and that improved version will improve itself, and on and on.

I wonder how you can claim that. It sounds like a religious belief.

https://en.wikipedia.org/wiki/Quantum_algorithm

Quote
Problems which are undecidable using classical computers remain undecidable using quantum computers. What makes quantum algorithms interesting is that they might be able to solve some problems faster than classical algorithms.

*

unreality

  • Starship Trooper
  • *******
  • 443
Re: GPU paralleling question
« Reply #10 on: November 18, 2017, 08:34:57 pm »
A quantum computer is when we will definitely get the Singularity, but hopefully it won't require that. The movie Automata comes to mind. It will absolutely positively 100% guarantee happen. It's only a matter of when. At least by my definition of what Singularity means, which is when AI will be smart enough to improve itself, and that improved version will improve itself, and on and on.

I wonder how you can claim that. It sounds like a religious belief.

https://en.wikipedia.org/wiki/Quantum_algorithm

Quote
Problems which are undecidable using classical computers remain undecidable using quantum computers. What makes quantum algorithms interesting is that they might be able to solve some problems faster than classical algorithms.

So you're one of those humans who refuse to believe AI will surpass us? To me that seems like human ego. Why do you call the obvious a religion? Take a look at the growth rate of science. Seems obvious to me.

*

infurl

  • Administrator
  • ***********
  • Eve
  • *
  • 1372
  • Humans will disappoint you.
    • Home Page
Re: GPU paralleling question
« Reply #11 on: November 18, 2017, 09:50:24 pm »
So you're one of those humans who refuse to believe AI will surpass us? To me that seems like human ego. Why do you call the obvious a religion? Take a look at the growth rate of science. Seems obvious to me.



*

unreality

  • Starship Trooper
  • *******
  • 443
Re: GPU paralleling question
« Reply #12 on: November 18, 2017, 09:55:02 pm »
So you're one of those humans who refuse to believe AI will surpass us? To me that seems like human ego. Why do you call the obvious a religion? Take a look at the growth rate of science. Seems obvious to me.


smh

*

ranch vermin

  • Not much time left.
  • Terminator
  • *********
  • 947
  • Its nearly time!
Re: GPU paralleling question
« Reply #13 on: November 19, 2017, 06:20:41 am »
because gpus go fast, you can end up being that lazy to the point its only 5 times faster.    but the performance is actually there if you can code it half decently.

*

ranch vermin

  • Not much time left.
  • Terminator
  • *********
  • 947
  • Its nearly time!
Re: GPU paralleling question
« Reply #14 on: November 19, 2017, 02:31:05 pm »
your factors are wrong,  because are u sure he wasnt using an old one?

You can actually put a filter framework on the cpu, how gpus do it, and it will be NONCOMPUTING basicly, in minutes per frame, when the gpu is finishing frame after frame under a second,  you can code raytracers to know that.

 


Requirements for functional equivalence to conscious processing?
by DaltonG (General AI Discussion)
November 19, 2024, 11:56:05 am
Will LLMs ever learn what is ... is?
by HS (Future of AI)
November 10, 2024, 06:28:10 pm
Who's the AI?
by frankinstien (Future of AI)
November 04, 2024, 05:45:05 am
Project Acuitas
by WriterOfMinds (General Project Discussion)
October 27, 2024, 09:17:10 pm
Ai improving AI
by infurl (AI Programming)
October 19, 2024, 03:43:29 am
Atronach's Eye
by WriterOfMinds (Home Made Robots)
October 13, 2024, 09:52:42 pm
Running local AI models
by spydaz (AI Programming)
October 07, 2024, 09:00:53 am
Hi IM BAA---AAACK!!
by MagnusWootton (Home Made Robots)
September 16, 2024, 09:49:10 pm
LLaMA2 Meta's chatbot released
by spydaz (AI News )
August 24, 2024, 02:58:36 pm
ollama and llama3
by spydaz (AI News )
August 24, 2024, 02:55:13 pm
AI controlled F-16, for real!
by frankinstien (AI News )
June 15, 2024, 05:40:28 am
Open AI GPT-4o - audio, vision, text combined reasoning
by MikeB (AI News )
May 14, 2024, 05:46:48 am
OpenAI Speech-to-Speech Reasoning Demo
by MikeB (AI News )
March 31, 2024, 01:00:53 pm
Say good-bye to GPUs...
by MikeB (AI News )
March 23, 2024, 09:23:52 am
Google Bard report
by ivan.moony (AI News )
February 14, 2024, 04:42:23 pm
Elon Musk's xAI Grok Chatbot
by MikeB (AI News )
December 11, 2023, 06:26:33 am

Users Online

394 Guests, 0 Users

Most Online Today: 597. Most Online Ever: 2369 (November 21, 2020, 04:08:13 pm)

Articles