Page 2 of 9 FirstFirst 12345 ... LastLast
Results 17 to 32 of 134

Thread: Nvidia announces Tesla P100 GPU with Pascal architecture

  1. #17
    Registered User
    Join Date
    Jan 2016
    Posts
    14
    Thanks
    0
    Thanked
    0 times in 0 posts

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    And the gamers hardware is last in business industry...

  2. #18
    Senior Member
    Join Date
    Dec 2013
    Posts
    3,526
    Thanks
    504
    Thanked
    468 times in 326 posts

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    Quote Originally Posted by scaryjim View Post
    OTOH, the announcement just keeps raising questions. Quite apart from the DP heavy design (which might not play out well in the gaming market), I want to know what they're doing with the HBM2 to only get 720GB/s out of it (and quite why they think that's 3x Maxwell when the 980 Ti and Titan X pull 336GB/s)? AMD's 4 stack HBM1 implementation returned 512GB/s theoretical bandwidth. The more I hear about Pascal the less impressive it appears to be....
    I've not done much research but could the lower throughput on the RAM be down to ECC?
    I think they maybe basing it on the M40's 288 GB/sec memory bandwidth.

  3. #19
    Senior Member
    Join Date
    Feb 2006
    Posts
    400
    Thanks
    9
    Thanked
    46 times in 39 posts
    • iworrall's system
      • Motherboard:
      • Asus B350M TUF Gaming
      • CPU:
      • AMD Ryzen 2700X
      • Memory:
      • 32GB C14 Team Group "Dark Pro"
      • Storage:
      • 512Gb Samsung 970 Evo NVMe
      • Graphics card(s):
      • Nvidia RTX3080 FE
      • PSU:
      • BeQuiet 1KW
      • Case:
      • Thermaltake Core V21
      • Operating System:
      • Windows 10 / Kubuntu 19.04
      • Monitor(s):
      • Acer Predator X34A
      • Internet:
      • Sky Fibre

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    Nvidia are probably saving any GeForce Pascal announcements for E3 in June.

  4. #20
    root Member DanceswithUnix's Avatar
    Join Date
    Jan 2006
    Location
    In the middle of a core dump
    Posts
    12,342
    Thanks
    714
    Thanked
    1,408 times in 1,190 posts
    • DanceswithUnix's system
      • Motherboard:
      • Asus X470-PRO
      • CPU:
      • 3700X
      • Memory:
      • 32GB 3200MHz ECC
      • Storage:
      • 1TB Linux, 1TB Games (Win 10)
      • Graphics card(s):
      • Asus Strix RX Vega 56
      • PSU:
      • 650W Corsair TX
      • Case:
      • Antec 300
      • Operating System:
      • Fedora 33 + Win 10 Pro 64 (yuk)
      • Monitor(s):
      • Benq XL2730Z 1440p + Iiyama 27" 1440p
      • Internet:
      • Zen 80Mb/20Mb VDSL

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    Quote Originally Posted by Corky34 View Post
    I've not done much research but could the lower throughput on the RAM be down to ECC?
    I think they maybe basing it on the M40's 288 GB/sec memory bandwidth.
    I wouldn't think so...

    ECC only really hurts if you want to just change a little bit of data, as you have to read in a line to recalculate the check code and write the whole line back. In something like a GPU which does big bursts of data, and is heavily cached I wouldn't expect ECC to really impact much at all as you are going to be dumping full width cache lines into RAM so don't have to do that slow read-modify-write cycle.

  5. #21
    Senior Member
    Join Date
    Dec 2013
    Posts
    3,526
    Thanks
    504
    Thanked
    468 times in 326 posts

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    My understanding of ECC is different then, or have i misunderstood?

    I thought EEC used an extra physical unit (RAM chip, maybe a HBM module?) to store the information (database?) of the parity bits, from what I've since read about ECC RAM it seems it doesn't reduce the throughput by any significant amount (1-2%), but i was thinking more along the lines of whether having ECC on a graphics card would reduce the available memory lanes, i.e with 4 HBM modules would you lose one of those 1024-bit memory lanes because you're using one module to store information of the parity bits on the other three modules?

    Quote Originally Posted by DanceswithUnix View Post
    Ah, so that's how they manage an announcement when they can't ship product, they say you can have one in June if you have $130000. Not many people will notice if that deadline slips by
    My understanding is that rather hefty price tag is for eight Tesla P100 boards and the supercomputer to go with it.
    Last edited by Corky34; 06-04-2016 at 01:39 PM.

  6. #22
    Banhammer in peace PeterB kalniel's Avatar
    Join Date
    Aug 2005
    Posts
    30,757
    Thanks
    1,789
    Thanked
    3,289 times in 2,647 posts
    • kalniel's system
      • Motherboard:
      • Gigabyte Z390 Aorus Ultra
      • CPU:
      • Intel i9 9900k
      • Memory:
      • 32GB DDR4 3200 CL16
      • Storage:
      • 1TB Samsung 970Evo+ NVMe
      • Graphics card(s):
      • nVidia GTX 1060 6GB
      • PSU:
      • Seasonic 600W
      • Case:
      • Cooler Master HAF 912
      • Operating System:
      • Win 10 Pro x64
      • Monitor(s):
      • Dell S2721DGF
      • Internet:
      • rubbish

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    Quote Originally Posted by Corky34 View Post
    My understanding is that rather hefty price tag is for eight Tesla P100 boards and the supercomputer to go with it.
    Indeed, but that's still the only way to 'get' one in June For the cards you have to wait for Q1 2017.
    Last edited by kalniel; 06-04-2016 at 02:06 PM.

  7. #23
    Not a good person scaryjim's Avatar
    Join Date
    Jan 2009
    Location
    Gateshead
    Posts
    15,196
    Thanks
    1,231
    Thanked
    2,291 times in 1,874 posts
    • scaryjim's system
      • Motherboard:
      • Dell Inspiron
      • CPU:
      • Core i5 8250U
      • Memory:
      • 2x 4GB DDR4 2666
      • Storage:
      • 128GB M.2 SSD + 1TB HDD
      • Graphics card(s):
      • Radeon R5 230
      • PSU:
      • Battery/Dell brick
      • Case:
      • Dell Inspiron 5570
      • Operating System:
      • Windows 10
      • Monitor(s):
      • 15" 1080p laptop panel

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    Quote Originally Posted by Corky34 View Post
    ... i was thinking more along the lines of whether having ECC on a graphics card would reduce the available memory lanes, i.e with 4 HBM modules would you lose one of those 1024-bit memory lanes because you're using one module to store information of the parity bits on the other three modules?
    It sounds like your mental image of how ECC works is much like a RAID5 array! I believe that might be true for some implementations, but I'd expect a graphics processor with limited numbers of memory chips available to have some form of in-chip error detection. Also, if an entire stack of HBM was used for parity the GPU would only have access to 12GB of RAM, not 16GB as nvidia claim.

  8. #24
    Oh Crumbs.... Biscuit's Avatar
    Join Date
    Feb 2007
    Location
    N. Yorkshire
    Posts
    11,193
    Thanks
    1,394
    Thanked
    1,091 times in 833 posts
    • Biscuit's system
      • Motherboard:
      • MSI B450M Mortar
      • CPU:
      • AMD 2700X (Be Quiet! Dark Rock 3)
      • Memory:
      • 16GB Patriot Viper 2 @ 3466MHz
      • Storage:
      • 500GB WD Black
      • Graphics card(s):
      • Sapphire R9 290X Vapor-X
      • PSU:
      • Seasonic Focus Gold 750W
      • Case:
      • Lian Li PC-V359
      • Operating System:
      • Windows 10 x64
      • Internet:
      • BT Infinity 80/20

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    Quote Originally Posted by DanceswithUnix View Post
    Ah, so that's how they manage an announcement when they can't ship product, they say you can have one in June if you have $130000. Not many people will notice if that deadline slips by
    Perhaps this is how much it will cost them to get any kind of priority ahead of AMD In the HBM2 production line.

  9. #25
    Registered+
    Join Date
    Oct 2011
    Posts
    46
    Thanks
    0
    Thanked
    1 time in 1 post

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    Quote Originally Posted by Xlucine View Post
    Hmm - going from the 28,672 cuda cores in the GDX-1, that works out to just 3,584 cores per GPU - that's surprisingly low. 15Bn transistors is almost double what a titan X has, yet a titan X has 3,072 cores. It's also a factor of 7, which is odd for a computer (3584=7*2^9). Perhaps this is a 4096 core GPU with 1/8th disabled for yields? I've heard the 16nm node is about twice as space efficient as the 28nm, so it's probably somewhere around a square inch (600ish mm^2) which is rather bold for a new process node - this makes more sense if they're fusing off an eighth to boost yields, although it doesn't explain why cuda cores take so many more transistors now.
    It has 60 SM's 56 activated in the GPU in the article. Each SM has 64 single-precision stream processors and 32 double-precision stream processors. So 3840 single-precision and 1920 double-precision in the full chip (3584 and 1792 in the article card) for comparison the last GPU Nvidia made with dedicated double-precision cores the GK110 in the GTX 780 Ti had 2880 single-precision cores and 960 double-precision cores.

  10. Received thanks from:

    Xlucine (07-04-2016)

  11. #26
    Registered+
    Join Date
    Oct 2011
    Posts
    46
    Thanks
    0
    Thanked
    1 time in 1 post

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    Quote Originally Posted by Platinum View Post
    10.6 tflops seems a little low, the Fury x is rated at 8.6tflops
    That's single precision in double precision the Tesla P100 has 5.6 TFLOPS compared to the Fury X's 537.6 GFLOPS.

  12. #27
    Senior Member
    Join Date
    Jun 2013
    Location
    ATLANTIS
    Posts
    1,207
    Thanks
    1
    Thanked
    28 times in 26 posts

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    I can't wait for Witcher 3 performance at 4K.......the GPU killer.

  13. #28
    Senior Member
    Join Date
    Dec 2013
    Posts
    3,526
    Thanks
    504
    Thanked
    468 times in 326 posts

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    Quote Originally Posted by kalniel View Post
    Indeed, but that's still the only way to 'get' one in June For the cards you have to wait for Q1 2017.
    True but even if you could buy one you couldn't use it as you'd need a MoBo with a NVLink connection, apparently Nvidia says the PCI Express bus doesn't provide enough bandwidth to keep the card busy.

    Quote Originally Posted by scaryjim View Post
    It sounds like your mental image of how ECC works is much like a RAID5 array! I believe that might be true for some implementations, but I'd expect a graphics processor with limited numbers of memory chips available to have some form of in-chip error detection. Also, if an entire stack of HBM was used for parity the GPU would only have access to 12GB of RAM, not 16GB as nvidia claim.
    The first thing that sprung to mind was ECC RAM and how each stick of ram will have an extra memory chip but i guess it maybe similar to a RAID5 array, I'm just guessing though so without further reading i couldn't say for certain how either Nvidia or AMD configure things when it comes to cards designed for GPU-accelerated computation, although comparing the M40's 288 GB/sec memory bandwidth with something like the Titan X's 336.5 GB/sec shows a similar disparity between their HPC and gaming cards.
    Last edited by Corky34; 06-04-2016 at 03:44 PM.

  14. #29
    Senior Member
    Join Date
    Mar 2010
    Posts
    2,567
    Thanks
    39
    Thanked
    179 times in 134 posts

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    where did it say Q1 2017 for retail cards?

  15. #30
    root Member DanceswithUnix's Avatar
    Join Date
    Jan 2006
    Location
    In the middle of a core dump
    Posts
    12,342
    Thanks
    714
    Thanked
    1,408 times in 1,190 posts
    • DanceswithUnix's system
      • Motherboard:
      • Asus X470-PRO
      • CPU:
      • 3700X
      • Memory:
      • 32GB 3200MHz ECC
      • Storage:
      • 1TB Linux, 1TB Games (Win 10)
      • Graphics card(s):
      • Asus Strix RX Vega 56
      • PSU:
      • 650W Corsair TX
      • Case:
      • Antec 300
      • Operating System:
      • Fedora 33 + Win 10 Pro 64 (yuk)
      • Monitor(s):
      • Benq XL2730Z 1440p + Iiyama 27" 1440p
      • Internet:
      • Zen 80Mb/20Mb VDSL

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    Quote Originally Posted by Corky34 View Post
    My understanding of ECC is different then, or have i misunderstood?

    I thought EEC used an extra physical unit (RAM chip, maybe a HBM module?) to store the information (database?) of the parity bits, from what I've since read about ECC RAM it seems it doesn't reduce the throughput by any significant amount (1-2%), but i was thinking more along the lines of whether having ECC on a graphics card would reduce the available memory lanes, i.e with 4 HBM modules would you lose one of those 1024-bit memory lanes because you're using one module to store information of the parity bits on the other three modules?
    A long time ago people put used parity on each byte of ram. That tells you if a single bit of ram gets flipped, but not which one so you can't correct it. When memory dimms got wide enough that a 64 bit DIMM with parity was 72 bit, some bright spark worked out that using a "hamming code" would tell you if *any* two bits were corrupted and if just one was corrupted can calculate which one so it can be flipped back.

    For that to work, you need to write out to ram in groups of 64 bits to calculate the extra 8 bits of ecc as all 64 bits are contributing. If you are going to write out 64 bits anyway, then it doesn't slow you down in the slightest which is why modern computers which are good at grouping writes together into big lumps aren't slowed down much.

    At a guess, I would say a Tesla card would be slower than a Geforce on the memory accesses just to keep the heat down. If you save 20W per card, then with 8 cards in a chassis that's 160W (numbers a pure guess, but you see what I mean that it adds up quick).

  16. Received thanks from:

    Corky34 (07-04-2016)

  17. #31
    Senior Member watercooled's Avatar
    Join Date
    Jan 2009
    Posts
    11,459
    Thanks
    1,539
    Thanked
    1,024 times in 868 posts

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    Quote Originally Posted by Corky34 View Post
    True but even if you could buy one you couldn't use it as you'd need a MoBo with a NVLink connection, apparently Nvidia says the PCI Express bus doesn't provide enough bandwidth to keep the card busy.
    The marketing slides show the GPU has PCIe interface - NVlink is a server interconnect, not a desktop feature. PCIe should be more than enough for a gaming workload, but even if for some reason Pascal was massively more bus-hungry than any previous card ever and PCI was suddenly not enough, tough, because you're getting a PCIe connection to the Intel/AMD CPU.


    But I agree with others, it looks very much like a HPC-focussed card rather than a gaming GPU - the size alone is likely to make it prohibitively expensive for quite some time. And then there are things like the NVlink interfaces which take an unknown size on the silicon, but if the non-gaming-applicable stuff isn't too large I guess they could just harvest dies where it's broken to sell as gaming cards.

    Edit: Just having read more of the thread, yeah a lot seems to be off with the announcement. Not least the absurd 'first' claims about CoWoS, etc. Also, what happened to that 10x performance claim? I mean we knew it was always going to be in some really specific mixed-precision workloads but I've not seen it mentioned again?

    Also, have we seen any evidence of working silicon yet? I only recall seeing renders? Though at least they're now actually showing HBM rather than what looked like HMC on the older ones! xD
    Last edited by watercooled; 06-04-2016 at 10:19 PM.

  18. #32
    Senior Member
    Join Date
    Aug 2009
    Location
    UK
    Posts
    431
    Thanks
    20
    Thanked
    33 times in 27 posts
    • Jace007's system
      • CPU:
      • Intel i7 7700k
      • Memory:
      • 16GB
      • Storage:
      • 500GB SSD
      • Graphics card(s):
      • nVidia 1080
      • PSU:
      • EVGA 750w
      • Operating System:
      • WinLOW

    Re: Nvidia announces Tesla P100 GPU with Pascal architecture

    OTOH, the announcement just keeps raising questions. Quite apart from the DP heavy design (which might not play out well in the gaming market), I want to know what they're doing with the HBM2 to only get 720GB/s out of it (and quite why they think that's 3x Maxwell when the 980 Ti and Titan X pull 336GB/s)? AMD's 4 stack HBM1 implementation returned 512GB/s theoretical bandwidth. The more I hear about Pascal the less impressive it appears to be....[/QUOTE]

    I found this. x3 the performance of Maxwell = Tesla M40 which is based on Maxwell.

    The Tesla P100 package includes four 4-Hi HBM2 stacks, for a total of 16 GB of memory, and 720 GB/s peak bandwidth. That’s three times as much bandwidth as Nvidia’s previous flagship accelerator the Tesla M40. The Telsa P100’s bandwidth figure is below that of the JEDEC HBM2 spec that SK Hynix & Samsung both adhere to. Which dictates that every 4-HI HBM2 stack should operate at a 2Ghz clock speed to deliver 256GB/s of bandwidth for a total of 1TB/s for four stacks. The HBM2 modules on the Tesla P100 actually operate well below the spec at only 1.4Ghz.
    This is part of Nvidia to reduce the overall power of the package.

Page 2 of 9 FirstFirst 12345 ... LastLast

Thread Information

Users Browsing this Thread

There are currently 1 users browsing this thread. (0 members and 1 guests)

Posting Permissions

  • You may not post new threads
  • You may not post replies
  • You may not post attachments
  • You may not edit your posts
  •