Become a MacRumors Supporter for $50/year with no ads, ability to filter front page stories, and private forums.

rittchard

macrumors 6502
Original poster
Aug 12, 2007
351
46
Model: 3.06GHz, GT 130, 1 TB
Bootcamp: Windows 7 Beta, 181.71 nVidia drivers

3DMark06 (default settings, 1280x1024) - 9645

3DVantage (default Performance settings) - P2643

Windows Experience Index - 5.9
- Processor 6.6
- Memory 5.9
- Graphics 7.9
- Gaming graphics 5.9
- Disk transfer 5.9

XBench - 201.39

From PC Magazine:

http://www.pcmag.com/image_popup/0,1871,iid=230029,00.asp
 
I am very curious about further details on this G130 graphic card as it seems to be a pretty good one... It makes sense if you look at the additional costs (40$) you have to pay for an ATI 4850. Maybe this Nvidia beast isn't that bad after all...
 
Can you please run GPU-Z and tell us what the video card is and at least the memory interface one it? :D

I tried running GPU-Z and didn't get valid info off of it.

I mentioned this elsewhere but it registers as a "Pre-release 9600GT". Most of the other detailed info didn't show up and some of the numbers that did show looked like errors.

Off of 3DVantage it registered as having 64 stream processors, 256-bit interface. Chipset was G94, rev. A1 I think. Also somewhere it's listed as PCI-E x16 ver 2, not sure what ver 2 is.

As to the other comment above, everything so far is pointing to this being a *desktop* version, not mobile - but perhaps an underclocked desktop version. The relatively lower graphics score on 3DVantage certainly seem to support the lower clock speed theory, but the score in the 2000s indicates it's probably not a mobile version.
 
Off of 3DVantage it registered as having 64 stream processors, 256-bit interface. Chipset was G94, rev. A1 I think. Also somewhere it's listed as PCI-E x16 ver 2, not sure what ver 2 is.

As to the other comment above, everything so far is pointing to this being a *desktop* version, not mobile - but perhaps an underclocked desktop version.

What is pointing to it being a desktop card? It looks a lot like a 9800M GS. When compared to the card in the new iMacs, (specs taken from your other thread) it has the same number of stream processors, the same core and shader clock speeds, the same memory speeds, and is also G94 (I am unsure which revision). This card seems to be the best fit, and is also a mobile card.
 
3DMark06 (default settings, 1280x1024) - 9645

Can you give us a breakdown of the 3 scores from 3DMark06? I'm curious of what the graphics scores were individually.

Edit: Never mind, found your breakdown in a different thread.
 
Is there any proof of rittchard's iMac getting those scores??? Can you post some screens,rittchard???
 
Isn't there an 9400 in all models of the new iMacs, since it being part of the chipset?

Can it be used together with GT120/130? Especially in windows with mature nvidian drivers? That would be swell!
 
Isn't there an 9400 in all models of the new iMacs, since it being part of the chipset?

Can it be used together with GT120/130? Especially in windows with mature nvidian drivers? That would be swell!

(1) Benchmarks seem a bit odd (unexpected); check out notebookcheck - it lists gt 130m for 6k on 3d mark 06... here you say 9k... interesting...

http://www.notebookcheck.net/NVIDIA-GeForce-GT-130M.13789.0.html

same as radeon hd4850. so is it a mobile version or desktop?

(2) good point about 9400; it must be there! it's unthinkable that they'd have a different design for models with discrete cards. The question is, can you use it?
 
I tried running GPU-Z and didn't get valid info off of it.

I mentioned this elsewhere but it registers as a "Pre-release 9600GT". Most of the other detailed info didn't show up and some of the numbers that did show looked like errors.

Off of 3DVantage it registered as having 64 stream processors, 256-bit interface. Chipset was G94, rev. A1 I think. Also somewhere it's listed as PCI-E x16 ver 2, not sure what ver 2 is.

As to the other comment above, everything so far is pointing to this being a *desktop* version, not mobile - but perhaps an underclocked desktop version. The relatively lower graphics score on 3DVantage certainly seem to support the lower clock speed theory, but the score in the 2000s indicates it's probably not a mobile version.

Those 3Dvantage numbers seem inconsistent with NVidia's specs for the GT 130 (M or not), this means that 3Dvantage is incorrect OR that the card is not really the GT 130:

http://www.nvidia.com/object/product_geforce_gt_130m_us.html
http://www.nvidia.com/object/product_geforce_gt_130_us.html

Is there a way to get the so called "GT 130"'s clock speed? The clock on the standard 9600GT is between 650-600MHz. EDIT: When I asked on a PC forum about reading the GPU clock from a video card, someone pointed me to Riva Tuner (I haven't ever used this FYI), perhaps you could use it tell us all the clock on it?: http://www.guru3d.com/index.php?page=rivatuner

I also want to note that when I called Apple, their "systems engineer" told me that the Radeon 4850 option was the Desktop version, so maybe lends credence to the GT 130 also being a Desktop chip.
 
Hey guys,

First off I have absolutely no reason to lie or make crap up, so you can take what I post here with a grain of salt or however you'd like. If anything, I'd just like to get to the bottom of what's in the machine myself, so I'm trying to provide whatever info I can and hoping someone will provide a definitive answer soon enough.

As far as benchmarks, they are what they are. I ran 3DVantage in all default settings (Performance) - definitely lower than what I'd expected but keep in mind I'm running Win 7 Beta with nVidia beta drivers. That said, a Vantage score in the 2500 range is not horrible; however, I'd expected something in the 3000-4000 range.

I ran both GPU-Z and Riva Tuner and thus far the info provided seems suspect. As I understand it these are all database oriented information providers, so all they will do is list info from a fixed database, based on some basic name bits read off the chip. If the database is not accurate, the info is worthless.

I'm posting below the report info Riva Tuner spit out. I don't know what it all means so someone smarter can try to decipher it :) The clock speeds all seem really low.

$ffffffffff ----------------------------------------------------------------
$ffffffffff Northbridge information
$ffffffffff ----------------------------------------------------------------
$0400000000 Description : unknown
$0400000001 Vendor ID : 10de (NVIDIA)
$0400000002 Device ID : 0a82
$0400000003 AGP bus : not supported
$ffffffffff ----------------------------------------------------------------
$ffffffffff Display adapter information
$ffffffffff ----------------------------------------------------------------
$0000000000 Description : NVIDIA GeForce 9600 GT (Prerelease - WDDM 1.1)
$0000000001 Vendor ID : 10de (NVIDIA)
$0000000002 Device ID : 062e
$0000000003 Location : bus 2, device 0, function 0
$0000000004 Bus type : PCIE
$000000000f PCIE link width : 16x supported, 1x selected
$0000000009 Base address 0 : d2000000 (memory range)
$000000000a Base address 1 : c0000000 (memory range)
$000000000b Base address 2 : none
$000000000c Base address 3 : d0000000 (memory range)
$000000000d Base address 4 : none
$000000000e Base address 5 : 00001000 (I/O range)
$ffffffffff ----------------------------------------------------------------
$ffffffffff NVIDIA specific display adapter information
$ffffffffff ----------------------------------------------------------------
$0100000000 Graphics core : G94 revision A1 (64sp)
$0100000001 Hardwired ID : 062e (ROM strapped to 062e)
$0100000002 Memory bus : 256-bit
$0100000003 Memory type : DDR3 (RAM configuration 00)
$0100000004 Memory amount : 524288KB
$0100000100 Core clock domain 0 : 168.750MHz
$0100000101 Core clock domain 1 : 337.500MHz
$0100000006 Memory clock : 100.000MHz (200.000MHz effective)
$0100000007 Reference clock : 25.000MHz/27.000MHz

And here is the System Info off nVidia control panel:

[Display]
Processor: Intel(R) Core(TM)2 Duo CPU E8435 @ 3.06GHz (3052 MHz)
Operating System: Windows 7 Ultimate, 64-bit
DirectX version: 10.0
GPU processor: GeForce GT 130
Driver version: 181.71
Stream processors: 64
Core clock: 529 MHz
Shader clock: 1323 MHz
Memory clock: 792 MHz (1584 MHz data rate)
Memory interface: 256-bit
Total available graphics memory: 2291 MB
Dedicated video memory: 512 MB
System video memory: 0 MB
Shared system memory: 1779 MB
Video BIOS version: 62.94.74.00.05
IRQ: 23
Bus: PCI Express x1 Gen2

Last but not least, off of GPU-Z:

ghx.png
 
At 600 MHz Core it looks like a 9800M GTS now. :eek:

I can't wait for a GPU-Z update to address this and get identifiers.

I wondered if this 130 wasn't the 9800m GTS myself. I'm edging ever closer to 99.9% sure that is what it this card will turn out to be. It is a decent card and make sense for an iMac, not only on being a decent card but being very energy efficient (75 Watt max). Still kind of a bummer though that this card is only an equivalent at the very best to the 8800 GS from the previous 3.06 Ghz iMacs. In fact NBCheck quotes it as being slightly slower (below) due to the lower clock rate. Why didn't Apple just make the 3.06Ghz machine with the ati and be done with it already....

BTW - anyone else notice the almost 2GB of "shared" RAM on those last tests?

source: notebookcheck

"Compared to desktop graphics cards, the 9800M GTS should be a bit slower than the GeForce 8800 GS (which has a higher memory clock rate)."

Manufacturer NVIDIA
Series GeForce 9000M
Codename NB9E-GT
Pipelines 64 - unified
Core Speed * 600 MHz
Shader Speed * 1500 MHz
Memory Speed * 800 MHz
Memory Bus Width 256 Bit
Memory Type GDDR3
Max. Amount of Memory 1024 MB
Shared Memory no
DirectX DirectX 10, Shader 4.0
Power Consumption 75 Watt
Transistors 504 Million
technology 55 / 65 nm
Features PureVideo HD, HybridPower,
Notebook Size large
Date of Announcement 15.07.2008
Information MXM 3
Link to Manufacturer Page http://www.nvidia.com/object/geforce_980...
* The specified clock rates are only guidelines for the manufacturer and can be altered by them.
 
The previous card in the iMacs at the high was the 8800M GTS. :D

I know I just bought one last week (3.06 Ghz). :D Was pondering returning it but with the 14 day clock ticking if I decided to swap the best I would be able to get would be a 130 unit. I figure I've got my 4 GB of Corsair SG RAM in this bad boy and all is screamin' fast. Considering the "upgrades" I think I'm going to stick right where I am. I think this will be more than I need until the next true push forward round of iMacs.
 
It's tough to suggest a new machine when for the most part the new GPU is a higher clocked 8800M GTS in 55nm.

Lets not get started on the godly 4850.
 
Register on MacRumors! This sidebar will go away, and you'll see fewer ads.