Whats old is new
So they have reinvented a form of DMA (Direct Memory Access), just with a different endpoint attached to it ?
Wasn't this problem solved in the 1970's ?
Nvidia, IBM, and university collaborators have a developed an architecture they say will provide fast fine-grain access to large amounts of data storage for GPU-accelerated applications, such as analytics and machine-learning training. Dubbed Big accelerator Memory, aka BaM, this is an interesting attempt to reduce the …
It's just an interesting way to get threads on GPU cores to talk direct to NVMe SSDs to get the data they need in a fine-grained, software-cached manner that specifically suits the access patterns of GPU-bound applications.
Yeah it involves DMA and all that. It's not claiming to have reinvented or come up with DMA; it's an application of it specific to GPU workloads.
Software cached just sounds like what IBM did with the Cell processor, where the data and code in the SPE's limited RAM was not controlled by a hardware cache mechanism second guessing what the software would ask for next, but by the orchestrating software putting data and code there for itself.
From the point of view of developers like me, that was fantastic. However the problem was that it was so far off the beaten track for a lot of developers it bombed as a concept and IBM dropped it, which was a real pity. Neat as this latest idea is, if it involves custom hardware on IBM's part I'd not be counting on them sustaining it as a dependable product line.
But it's in interesting idea. People have been looking for GPUs that can do more. Trouble is that, taken too far, you end up with something like an AMD APU....
Yeah, Z80s had a DMA mode where a system chip (usually the GFX chip) could set the CPU to DMA mode so that it could access main memory without the CPU cycles getting in the way
I think zilog had a FDD/HDD controller that used that mode too with a similer method to that in the figure of sending the controller the addresses to use along with block size, and the controller would use a DMA mode when the data had arrived from the disk and write it to memory.
I guess the reason this is really taking off is that the storage is no longer the bottleneck. In the past you needed something like a CPU to moderate access to storage to ensure no one device or process hogged storage. Now storage is no longer the bottleneck and everything can talk to it direct.
This post has been deleted by a moderator
A large part of the AMD implementation was having faster-than-PCIe hardware interconnects twixt the SSDs and the GPU chip. The SSDs would be seen by the GPU as a slower tier of memory in addition to its RAM. The nVidia solution involves off the shelf hardware with custom software and drivers.
Are both ideas trying to make things faster? Well yeah, thats usually the game!
The future if NVMe to GPU data transfer looks rosy, be it for computation or gamers.
This promises to gamers a future for loading textures and other graphics artifacts from NVMe to GPU with lower CPU overhead.
But still needs GPU vendors to provide support at their end both in revised hardware and drivers to achieve the objectives.
So this ElReg article appears to mimic for Linux what Microsoft have had in beta over past 18 months.
Here is to an open source and spec future. Hip hip..
Less than a week after IBM was ordered in an age discrimination lawsuit to produce internal emails in which its former CEO and former SVP of human resources discuss reducing the number of older workers, the IT giant chose to settle the case for an undisclosed sum rather than proceed to trial next month.
The order, issued on June 9, in Schenfeld v. IBM, describes Exhibit 10, which "contains emails that discuss the effort taken by IBM to increase the number of 'millennial' employees."
Plaintiff Eugene Schenfeld, who worked as an IBM research scientist when current CEO Arvind Krishna ran IBM's research group, sued IBM for age discrimination in November, 2018. His claim is one of many that followed a March 2018 report by ProPublica and Mother Jones about a concerted effort to de-age IBM and a 2020 finding by the US Equal Employment Opportunity Commission (EEOC) that IBM executives had directed managers to get rid of older workers to make room for younger ones.
Interview 2023 is shaping up to become a big year for Arm-based server chips, and a significant part of this drive will come from Nvidia, which appears steadfast in its belief in the future of Arm, even if it can't own the company.
Several system vendors are expected to push out servers next year that will use Nvidia's new Arm-based chips. These consist of the Grace Superchip, which combines two of Nvidia's Grace CPUs, and the Grace-Hopper Superchip, which brings together one Grace CPU with one Hopper GPU.
The vendors lining up servers include American companies like Dell Technologies, HPE and Supermicro, as well Lenovo in Hong Kong, Inspur in China, plus ASUS, Foxconn, Gigabyte, and Wiwynn in Taiwan are also on board. The servers will target application areas where high performance is key: AI training and inference, high-performance computing, digital twins, and cloud gaming and graphics.
Lenovo has unveiled a small desktop workstation in a new physical format that's smaller than previous compact designs, but which it claims still has the type of performance professional users require.
Available from the end of this month, the ThinkStation P360 Ultra comes in a chassis that is less than 4 liters in total volume, but packs in 12th Gen Intel Core processors – that's the latest Alder Lake generation with up to 16 cores, but not the Xeon chips that we would expect to see in a workstation – and an Nvidia RTX A5000 GPU.
Other specifications include up to 128GB of DDR5 memory, two PCIe 4.0 slots, up to 8TB of storage using plug-in M.2 cards, plus dual Ethernet and Thunderbolt 4 ports, and support for up to eight displays, the latter of which will please many professional users. Pricing is expected to start at $1,299 in the US.
Early details of the specifications for PCIe 7.0 are out, and it's expected to deliver data rates of up to 512 GBps bi-directionally for data-intensive applications such as 800G Ethernet.
The announcement from the The Peripheral Component Interconnect Special Interest Group (PCI SIG) was made to coincide with its Developers Conference 2022, held at the Santa Clara Convention Center in California this week. It also marks the 30th anniversary of the PCI-SIG itself.
While the completed specifications for PCIe 6.0 were only released this January, PCIe 7.0 looks to double the bandwidth of the high-speed interconnect yet again from a raw bit rate of 64 GTps to 128 GTps, and bi-directional speeds of up to 512 GBps in a x16 configuration.
IBM has quietly announced its first-ever cloudy mainframes will go live on June 30.
Big Blue in February disclosed its plans to provide cloud-hosted virtual machines running the z/OS that powers its mainframes. These would be first offered in a closed "experimental" beta under the IBM Wazi as-a-service brand. That announcement promised "on-demand access to z/OS, available as needed for development and test" with general availability expected "in 2H 2022."
The IT giant has now slipped out an advisory that reveals a “planned availability date” of June 30.
Arm is beefing up its role in the rapidly-evolving (yet long-standing) hardware-based real-time ray tracing arena.
The company revealed on Tuesday that it will introduce the feature in its new flagship Immortalis-G715 GPU design for smartphones, promising to deliver graphics in mobile games that realistically recreate the way light interacts with objects.
Arm is promoting the Immortalis-G715 as its best mobile GPU design yet, claiming that it will provide 15 percent faster performance and 15 percent better energy efficiency compared to the currently available Mali-G710.
Siemens and Nvidia don’t want manufacturers to imagine what the future will hold – they want to build a fancy digital twin that helps them to make predictions about whatever comes next.
During a press conference this week, Siemens CEO Roland Busch painted a picture of a future in which manufacturers are besieged with productivity, labor, and supply chain disruptions.
"The answer to all of these challenges is technology and digitalization," he said. "The point is, we have to make the digital twin as realistic as possible and bring it as close as possible to the real world."
Comment Facebook parent Meta has reportedly said it needs to increase its fleet of datacenter GPUs fivefold to help it compete against short-form video app and perennial security concern TikTok.
The oft-controversial tech giant needs these hardware accelerators in its servers by the end of the year to power its so-called discovery engine that will become the center of future social media efforts, according to an internal memo seen by Reuters that was written by Meta Chief Product Officer Chris Cox.
Separately, CEO Mark Zuckerberg told Meta staff on Thursday in a weekly Q&A the biz had planned to hire 10,000 engineers this year, and this has now been cut to between 6,000 and 7,000 in the shadow of an economic downturn. He also said some open positions would be removed, and pressure will be placed on the performance of those staying at the corporation.
Updated ERP vendor Infor is to end development of an on-premises and containerized version of its core product for customers running on IBM iSeries mid-range systems.
Born from a cross-breeding of ERP stalwarts Baan and Lawson, Infor was developing an on-premises containerized version of M3, dubbed CM3, to help ease migration for IBM hardware customers and offer them options other than lifting and shifting to the cloud.
Under the plans, Infor said it would continue to to run the database component on IBM i (Power and I operating system, formerly known as iSeries) while supporting the application component of the product in a Linux or Windows container on Kubernetes.
Updated In one of the many ongoing age discrimination lawsuits against IBM, Big Blue has been ordered to produce internal emails in which former CEO Ginny Rometty and former SVP of Human Resources Diane Gherson discuss efforts to get rid of older employees.
IBM as recently as February denied any "systemic age discrimination" ever occurred at the mainframe giant, despite the August 31, 2020 finding by the US Equal Employment Opportunity Commission (EEOC) that "top-down messaging from IBM’s highest ranks directing managers to engage in an aggressive approach to significantly reduce the headcount of older workers to make room for Early Professional Hires."
The court's description of these emails between executives further contradicts IBM's assertions and supports claims of age discrimination raised by a 2018 report from ProPublica and Mother Jones, by other sources prior to that, and by numerous lawsuits.
Biting the hand that feeds IT © 1998–2022