* Posts by HuBo

1115 publicly visible posts • joined 20 Nov 2023

Page:

Silicon photonics won’t matter ‘anytime soon’ says Broadcom CEO

HuBo Silver badge
Pint

Re: This thing we're behind the competition in

Exactly my thoughts! I mean, Google TPUs (developed with Broadcom and MediaTek) are all about optical interconnects, and are attracting Meta's interest (among others?). Nvidia's moving swiftly into photonics and Broadcom's retorting with photonic Tomahawks ... and, in other News, "According to Broadcom, its CPO tech is more than 3.5x more efficient than pluggables" ...

So, has Hock Tan lost touch with his company's efforts in this, or were these previous headlines meant to be humoristic; what gives?

Bezos-backed Unconventional AI aims to make datacenter power problems go away

HuBo Silver badge
Pint

Quite nice

Dylan's 2022 TFA showed a cool plot of the enhanced power efficiency of neuromorphics compared to CPUs and GPUs. I'm glad Unconvential AI is getting funding to pursue this sort of event-based analog spike-processing, or whatever it is they're not telling.

There's the ARM-based SpiNNaker out there, shiitake memristors, sub-threshold transistors, the Semantic Pointer Architecture (SPA) for building a brain, whatever this is, and what have you ... plenty of concepts and room to explore. I love it!

Brains neither run at GHz speeds nor consume GW of power so if GPUs "simulating" them do, they're likely doing it wrong ...

Amazon keeps the pressure on Intel, AMD with 192-core Graviton5 CPU

HuBo Silver badge
Windows

Very nice

Great to see Graviton5 supporting PCIe 6 out of the box. Hopefully that means we soon get to see how CXL 3.0 performs in real life with its multi-level switching, coherency, peer-to-peer DMA, and memory sharing -- that should ease the task of getting multiple sockets (or computational nodes) acting rather seamlessly as a large NUMA machine, when needed.

These should then be a great target for Virtual Fugaku if they also sport solid vector units.

India has satisfied its supercomputing needs, but not its ambitions

HuBo Silver badge
Windows

Neat

We've known since Fugaku that CPU-only HPC systems can do very well at 400+ PF/s on HPL, and leading perf on HPCG (until this month's Top500), so the AUM Neoverse V2 should be plenty fine imho, even without a GPU (just needs hefty vector units). It's not as energy efficient as a hybrid CPU+GPU system but way less demanding than the GigaWatt+ systems being considered for AI (so called) training and inference these days.

Hopefully the EU can also field itself some CPU-only 100+ PF/s systems running on Rhea 1 or 2 in the near future ...

Copackaged optics have officially found their killer app - of course it's AI

HuBo Silver badge
Windows

Illuminating

Great to see this focus on CPO and related advances imho. AWS saw the light last year and showed a great before-and-after picture of how it vastly cleaned up its rack designs (the beforelast photo). Google's also known for the reconfigurable optical interconnects used around its TPUs ...

One can send multiple signals (at different wavelengths) simultaneously, for example 128+ of them using a single ChromX laser, and do so in both directions (BiDi) at once, on a single fiber, without a need to wrap the cable in grounded foil thanks to inherent galvanic isolation. So, the throughput can be formidable, even with very few physical links.

It's super that Nvidia, Broadcom, Ayar, and even AMD (eg. Enosemi purchase) are getting to the point of deploying this tech in the field, finally! Quite notable also that Lightmatter's Passage is a reconfigurable optical interposer (iirc) which should help orchestrate throughput in response to workload, if needed. It's been a long wait ...

Nvidia's green500 dominance continues as France's Kairos super takes efficiency title

HuBo Silver badge
Holmes

Re: Interesting

... and, when the good folks at AMD, HPE Cray, and DOE/NNSA/LLNL get El Capitan to Rmax at 2.1 EF/s (i.e. 74% of it Rpeak), while maintaining its 60.9 GF/W (improving networking and/or orchestration, running at 34 MW) then it will fall on the same pareto line as Jupiter:

Leading GF/W = 85.6 - 7.44 * log10(Rmax)
The Octave/MATLAB code to add that to the above is (with a red star for the up-tuned Capitan):
hold on; semilogx(2089,60.9,'r*',logspace(0,3.5,21),85.6-7.44*linspace(0,3.5,21),'b:'); hold off
(see also the deep dive and comments on this at TNP)

HuBo Silver badge
Holmes

Interesting

With a machine's Rmax having quite an impact on GF/W (as noted in TFA), I'd submit that energy performance should be better than the following to qualify as remarkable (at present): Expected GF/W = 78.3 - 7.44 * log10(Rmax)

The two supercomputers that are clearly above that curve are Jupiter and El Capitan, but I'd expect upcoming Alice Recoque, Discovery, and other ExaFloppers to also be there (hopefully). If you have GNU Octave or MATLAB, you can check the situation with this script for Green500 systems beween 55 and 73 GF/W:

nvdrmx=[3.05 9.86 6.75 7.42 4.66 24.1 3.82 19.1 2.88 1000 3.12 123. 435. 5.04 13.2 98.5 3.14 10.2 6.42];

nvdeff=[73.3 70.9 69.4 68.8 68.2 68.0 67.2 66.9 65.4 63.3 63.0 62.0 61.0 61.0 60.3 59.3 58.2 57.0 55.6];

amdrmx=[2.99 24.5 31.7 31.1 24.4 19.2 208. 68.0 1809 15.9 9.89 46.1 27.2 478. 1353];

amdeff=[66.5 66.3 64.6 64.0 62.8 62.7 61.4 61.3 60.9 59.2 59.0 58.0 57.0 56.5 55.0];

semilogx(nvdrmx,nvdeff,'go',amdrmx,amdeff,'ro',logspace(0,3.3,21),78.3-7.44*log10(logspace(0,3.3,21)),'k:');xlabel('Rmax (PF/s)');ylabel('GF/W'); legend('Nvidia','AMD','expected')

SC25 gets heavy with mega power and cooling solutions

HuBo Silver badge
Windows

Yay, nay, or ouch?

It's interesting but unfortunate (imho) that the current AI hype-bubble is leading to this focus on high-density electricity production like this, especially through thermal techs. I can't help but think that what is truly needed are better ways to extract controlled electron flows directly from the otherwise disordered inner-state of matter, possibly through new metamaterials that act as one-way valves, or diodes, at the level of elemental pseudo-particles/waves (quantum or not). Maybe harvesting nuclear radiation through photovoltaics could work there to some extent (or somesuch)?

The days of LLM tech (with its wasteful energy consumption) as lead AI prospect are probably counted as well. For example, it seems from Thomas Hubert and team (a name almost as nice as "Bert Hubert") that Gemini's AlphaProof's relative success at Math Olympiads is mostly thanks to its use of classical AI, namely using L∃∀N (for formal mathematical reasoning) and tree search (some version of the A* algorithm?) to do its do, coupled with whatever Test-Time Reinforcement Learning (TTRL) is. It may match Gary Marcus's "neurosymbolic techniques" perspectives as well as that of DARPA's Shafto, and clearly doesn't work at all without the classical AI part, period.

Accordingly, datacenter that "will exceed 400,000" GPUs sound like a huge waste of resources if their focus will be on running LLMs. If they consume 150x what El Capitan does and yet don't produce 150x the computational oomph (at proper FP64 for HPC, and INT64 for classical AI) then they are a huge waste, full stop. A proper 150x El Capitan would crank 300 FP64 ExaFLOPs, which with MxP may result in 3.0 ZettaFLOPs of performance, and finally allow for high-resolution climate simulations at Earth-scale (among others). Granted the ICON team received the Gordon Bell Prize for climate modelling yesterday for its "Computing the Full Earth System at 1 km Resolution" on JEDI, Alps, and Jupiter, but using other physically-based models, or enhancing that resolution further (eg. to predict traveling wave derechos and traveling swirl tornados), still mandates Zettascale computing (iiuc).

Oh, and (almost unrelated) the other Gordon Bell Prize this year is for the Tsunami prediction research covered here back in August by Tobias ... cool stuff (imho)!

Eviden set to build France's first exascale supercomputer with AMD at the wheel

HuBo Silver badge
Pint

Way to go

Good to see Eviden progressing on its 2ⁿᵈ ExaFLOP system. The 12 MW looks ambitious at first but seeing how it is 2/5ᵗʰ of El Capitan's power draw it should lead to at least 2/5ᵗʰ its Rpeak, so 1.12+ EF/s. Some 15% generational perf improvement coupled with 78% efficiency (from the faster BXI?) could then reasonably give it an Rmax of 1.0 EF/s (or 20% on perf and 74% eff, ...).

What would be truly awesome though is if we could get 1 EF/s in 1 MW (or 1 PF/s in 1 kW) in the near-ish future (in full-fledged FP64)!

Scientific computing is about to get a massive injection of AI

HuBo Silver badge
Holmes

Blackwell: "high-performance linpack declined [but] (HPCG) benchmark – rose"

Good point. I guess that means we can approximate CHIE-4's HPCG/HPL ratio (Xeon+Blackwell) from ABCI 3.0's (Xeon+H200) with those figures (and/or vice-versa) ... say: (2.446/145)*(45/34)/(45/67) = 3.3% to CHIE-4's actual 2.8% ... perty close (rounding to 3%)!

Europe joins US as exascale superpower after Jupiter clinches Top500 run

HuBo Silver badge
Windows

Nice

Good to see Jupiter hitting 1.00 EF/s in 16 MW. That makes it #14 in Green500 for power efficiency, which is the first position in Exascalers (eg. El Capitan is #23). That's the right direction to go towards.

CHIE-4 is also interesting as the highest performing Xeon+B200 system so far. Its 24x MxP speedup is remarkable and hopefully won't be revised down as Tsubame4 was -- from 25x in Nov. '24 to a still respectable 16x in June '25. It's also neat that its HPCG perf is around 3% of its oomph on HPL, which is close to the ratio on CPU-only systems (Fugaku, Crossroads), better than other CPU+GPU machines (typ. near 1%), but still below vector engines (AOBA-S, TSUBASA, near 5%). If that's due to Blackwell (GPUDirect RDMA enhancements?) then it's great design, and I can't wait to see what the MI430X riposte will do!

Baidu answers China's call for home-grown silicon with custom AI accelerators

HuBo Silver badge
Windows

All tease and no strip

Hmmm ... I'd say the press releases on Baidu's Kunlun M100+ baked inference chips are rather thin-&-light on details (internal arch, networking, TDP, perf, litho, chip photo, die shot, supporting software stack ...). And if the 2 trillion (2-T) parms ERNIE 5.0 is out now, it must have been trained on some other hardware, maybe some Ascend 910 (as found in CloudMatrix 384) similar to how Huawei's 1-T PanGu-Σ was trained ... or some pre-sanctions Nvidia kit? More qestions than answers here ...

As for questions, they also "announced" (TFA link) Famou: "the World's First Commercially Available Self-evolving Agent", "able to quickly abstract complex problems and iterate automatically as conditions change" ... what!!?? Details and examples would be welcome on their part on this given the boldness of the claims.

Speaking of 1-T parameters though, Argonne's Ian Foster (and team) seems to think "an AI-native Scientific Discovery Platform (SDP) that connects models to tools, data, HPC, and robotics" based on "science-tuned foundation models" of that size could be worthwhile. I have my doubts. But he'll present that concept next Friday at SC25 (and there's an ArXiv on this from last year) ... might be worth a gander (or not?).

HPE details Vera Rubin blades for next-gen Cray supercomputers

HuBo Silver badge
Windows

Yummy

Looking forward to seeing more details on MI430X performance, with the hope of a cool 1.0+ FP64 PetaFlop/s in 1 kW for example (like a Roadrunner on a chip, but 1000x more power efficient).

Wonder what kind of new news there'll be on this in St.Louis at SC25 (Sunday to Friday)?

Microsoft teases agents that become ‘independent users within the workforce’

HuBo Silver badge
Pint

Re: Automate HR....

Cool analogy! Had to look it up and found it in the French WikiPedia (bio-robots), and the German WikiPedia (Bioroboter), but not in the English WikiPedia for some reason ... but it's at the NIH's NLM as biorobots, and there's a nice picture from Swiss (Cheese) media here.

Reminds me of Ludwig Von 88's song Cs 137 that goes "Nous les gars de Pripiat ..." (We, the guys from Pripyat ...). The efficiency and progress (ours once more) of this Agentic AI (so-called) virtual workfarce may well just lead to similar devastation ... with big beautiful workplace buildings empty of life until it's time for post-disaster decontamination ... by us, the disposable bio-robots ... imho!

MIT Sloan quietly shelves AI ransomware study after researcher calls BS

HuBo Silver badge
Pint

Refreshing

I love Beaumont and Hutchins' takes on this, which I'd summarize as: it's absurd jaw droppingly bad corporate marketing bozos cyberslop nonsense ... just rolls off the tongue!

Great to see outlandish AI claims being taken down a notch this way.

There's mushroom for improvement in fungal computing

HuBo Silver badge
Thumb Up

Love it

Seeing how shiitake essentially consist of a whole bunch of memristors that are fundamental to low-power neuromorphic computing, makes me wonder if thin-slicing them and inserting the result in a sandwich of BCI electrodes (grid arrays as found in a toaster, kind off) could result in interesting learning or stimulus processing abilities? Or would it be necessary to grow the shiitakes in the shape of a cauliflower first, possibly through transgenic hybridization (by gene gun? Like oyster mushrooms?)? Would the shroom's 36,000 sexes get in the way? Could these then be used for brain transplants?

The possibilities seem endless ... ;) (not to mention the delicious recipes!)

This is Doom, running headless, on Ubuntu Arm… on a satellite

HuBo Silver badge
Pint

Re: I'm still confused, 14 years later...

Good point! I guess if 1992 was year 1 of id Doom development, 1993 was year 2, and November 2011 became the start of year 20 ...

But hey, cool TFA and video (well worth the 8 minutes)! And as Ólafur notes, it should be interesting to see what folks end up doing with OPS-SAT VOLT (expected to launch in 2026) that'll feature the Leopard DPU with its Zynq UltraScale FPGA sporting quad Cortex-A53 CPUs (as compared to the Cyclone V SoC's dual-core Cortex-A9 in this past OPS-SAT).

Should be fun to watch ...

Berkeley boffins build better load balancing algo with AI

HuBo Silver badge
Windows

A bit surprised

In their paper, the UC Berkeley crew writes:

"OpenEvolve independently rediscovered and fully exploited a tensorized zigzag partitioning scheme, yielding an evolved EPLB algorithm that achieves a 5.0× speedup"

suggesting to me that the speedup technique was already known, but not necessarily applied to this specific problem(?). It also seems notable that in their Table 1, a lot of OpenEvolve applications resulted in 7% speedups (only), with "Adaptive Weight Compression" being 14% worse instead than without AI "optimization" ... maybe they'll want to tone down the "amazing" "never seen before" "5x" speedup bit a little ...

Despite this though, I wouldn't be against seeing ADRS applied to improving on reverse Cuthill-McKee algorithms, multi-frontal methods, Delaunay tetrahedralization with matching faces, ADI over simplices (NOT rectangles), and the likes, especially in the context of large multi-CPU systems where the type of EPLB they investigated is important. Could be interesting for reconfigurable Maverick-2-type systems too!

Bottom line, if this tech can help tame the CPU memory problem and make the Pain of Parallel Programming more bearable, then I'm all for it. Hopefully they get tested over properly serious challenges though, rather than AI nombrilism ... imho.

Don't take AI to Thanksgiving: Bots have hidden biases

HuBo Silver badge
Pint

Re: Sussman attains enlightenment

¡Moscas gratas muynacho! Your kōan-foo carves me bar just right ... Zen up++! ;)

NextSilicon Maverick-2 promises to blow away the HPC market Nvidia left behind

HuBo Silver badge
Windows

Much needed

Yeah, the good folks at Sandia Vanguard describe the Mavericks as a runtime-reconfigurable accelerator which vastly helps it adapt its dataflow to workload specifics ... very neat! We badly need this capability also in scale-up/out networking to propagate the benefits of this flexibility to the system scale (with PCIe 6, CXL 3, and CPO).

The 600 GF/s FP64 perf on HPCG may sound low compared to 45 TF/s dense (eg. HPL) on a GB200, but checking with Top500 shows that the HPCG perf of Frontier (MI250X), Aurora (GPU Max), and Alps (GH200) is less than 1% of their perf on dense HPL (aka Rmax). In other words, it would take a 60+ TF/s FP64 GPU to get the 600 GF/s on HPCG that Maverick-2 (750W dual-die) gets. Interestingly, TNP reports (linked under "pointed out") that this Maverick cranks 40 TF/s on dense calcs, making its HPCG oomph 1.5% of its dense grunt, which is 1.5x to 3x better than seen in current Top500 GPUs ... nice!

The 2023 Gordon Bell Prize for Climate Modelling rewarded the SCREAM team for their pioneering exascale 1.26 simulated years per day of cloud-resolving earth atmosphere simulation at 3.25 km resolution. Getting the resolution down to 1 km should require (3.25 x 3.25 x 3.25)² more computations (approx. 1000x, i.e. Zettaflopping) and any tech that helps us get there efficiently is welcome imho (eg. Maverick-2). Meanwhile, some folks claim they can compute the Full Earth System at 1km, with 91.8 simulated days per day (1/4-year per day), on Alps and Jupiter, which should be interesting to see at SC25, if it works (software and hardware combined to improve perf further than either alone?)!

AI boffins teach office supplies to predict your next move

HuBo Silver badge
Holmes

Re: Teach the objects to sing...

Yeah ... reminds me a bit more of skutters though, as you can see with the moving trivets, shaking keycup, and usb-plug zamboni in the lab's 30-second youtube "Towards Unobtrusive Physical AI" and their 4:46 video to the right of that which has the stapler-moving unit near 4:20 (mostly harmless except for the knife shuffling at 3:00 and 3:20 -- suspense!).

It's cool as experimentation imho (research), especially the 30-sec youtube of "Constraint-Driven Robotic Surfaces" (on the same lab page) that has a shape-shifting wall that automatically re-multi-purposes itself in response to perceived intent by whomever is there ... (I'd say it could at least be part of a fun futuristic movie!). Interesting angles imho ...

The $100B memory war: Inside the battle for AI's future

HuBo Silver badge
Windows

To HBM or not to LPDDR5x, and/or MR DIMM?

HBM might be a bit of a stopgap measure in this tensor-matrix-vector compute tech that underlies today's LLMs and related AI (so-called). Long-term, it should be best to consider in-memory compute and dataflow archs that are best suited to this kind of workload that involves very little case analysis and branching (dispatch) but consist instead of a buttload of multiplications and additions performed on hefty datasets held in memory. Distributing a whole bunch of weak-and-simple compute units throughout memory makes the most sense for this imho.

As an almost unrelated anecdote, I was comparing perf of an ILR (graph) interpreter and a JIT on some microcontrollers, with the JIT being generally 3x faster than the interpreter ... except on a board with a 600 MHz 32-bit MCU (Cortex-M7) hooked to 16-bit wide 150 MHz external RAM. For this board with 8x speed ratio (for 32-bit data), JIT code performed no faster than ILR interpretation. Analyzing the situation showed that the JIT process specialized the code such that case dispatches were removed (no longer needed) leaving mostly memory accesses and unconditional branches to be executed (with adds and mults in-between). The ratio of memory access to compute thereby increased relative to ILR interpretation, highlighting the severely limited speed of attached RAM (maybe that's what's going on with Python's JIT too?). In this case, reducing MCU clock to 150 MHz made the JIT 4x faster than ILR (at this same lower clock), but one would have obviously preferred for the RAM to be 4x faster instead and properly take advantage of the MCU's 600 MHz capability (premium).

Anyways, as stressed in TFA RAM speed is important (get the fastest you can), but also arch. For AI, getting compute units and mem close to one another, and distributed if possible, should help. Some other workloads may however benefit more from a graph processing beast arch ... among others, imho!

18 zettaFLOPS of new AI compute coming online from Oracle late next year

HuBo Silver badge
Windows

ZettaFLOPS shmzettaflops.

Wow! With SC25 (Nov 16–21) St. Louis, MO (not quite nicknamed "Chess capital lion of the gateway valley mound to the West World") right around the corner, I can't help thinking that if FP64 oomph scaled like FP4-to-TF32 in those GPUs (making it about 2 PF/s), an 800,000 GPU system would crank 1.6 ZettaFlops of HPC-appropriate horsepower (very much needed for very high-resolution whole-earth climate modeling)!

Instead of that, at only 80 TF/s per GPU in FP64, such system may put out just 64 Exaflops/s (EF/s) of compute (if they're all efficiently linked together) ... pfaaah! Even MxP might only raise this to a measly 640 EF/s of useful number crunching ... should I really have to get myself out of bed for this?! </he-he-he!>

DGX Spark, Nvidia’s tiniest supercomputer, tackles large models at solid speeds

HuBo Silver badge
Windows

Great review

Very thorough and complete. It'd be great if the usual suspects would send the ElReg Review Bureau a set of their Thor, M5 Max and Strix Halo units for a further comparative analysis imho!

It looks like the 128GB of RAM and 500 TF/s of dense oomph at FP4 (1 PF/s sparse) are the key features of this very cute desk machine, about the size of a Mac Mini of some generation. The Thor might be twice as fast on models of the same size (if MEM bandwidth is ok) but it is grey-ugly by comparison. And it seems there's no real purpose running LLMs of 3 billion parameters and smaller locally on Core i3 Whiskey Lakes with no usable GPU, so this Spark's ability to fine-tune 70B models, and run 120B models, could prove useful to some folks (eg. workers who's boss insist they use AI on the job, retirees endeavoring to stay in touch with tech, etc ...). But of course, they'll need to fix the Firefox out-of-memory issue first!

This unit reminds me that AMD should have put out a ½-, ⅓-, or ¼-scale MI300A, with added FP4 support, for exactly this space. Seems to me it could be solid competition to the GB10/Spark, and maybe help hook kids in early and often, cementing their favoring of the ROCm SOCm ecosystem and suchlikes! The competition should also help make such devices more affordable, and hopefully they'd sustain solid FP64 perf so that proper grad students involved in HPC might make good use of them too!

OpenAI GPT-5: great taste, less filling, now with 30% less bias

HuBo Silver badge
Gimp

Well, I think the outsized cognitive dissonance of the extreme right-wing is such that they label other extreme right-wingers that are of different genders, skin tone, or religion, as extreme left-wingers (cracks me right up!). But the jury's been out with a 100-page report and verdict on this for some time:

"jihadists too are extreme right-wing actors even if they are rarely referred to in such terms."

Then again, extreme right-wing mass murderers, like Marc Lépine (killed 14, in 1989, in Canada), Timothy McVeigh (killed 168, in 1995), Anders Breivik (killed 77, in 2011, in Norway), Dylann Roof (killed 9, in 2015), Alexandre Bissonnette (killed 6, in 2017, in Canada), Brenton Tarrant (killed 51, in 2019, in New Zealand), haven't really needed the help of Osama Bin Laden (killed 2,977, in 2001) to get noticed for their insane horrors.

The CSIS analysis of Left-Wing Terrorism and Political Violence in the United States over the past 10 years excludes those (by date or location) but still notes:

"Since 2020, only two fatalities have resulted from left-wing terrorist attacks in the United States: Luigi Mangione’s assassination of UnitedHealthcare CEO Brian Thompson in New York City in December 2024 and Michael Reinoehl’s fatal shooting of right-wing protester Aaron Danielson in Portland, Oregon, in August 2020 (if the Kirk killing is included, as seems likely, it would be a third fatality). Right-wing and jihadist attacks, by contrast, have caused far higher fatalities."

The data in their Table 1 drives the point home further imho, in the past 10 years, in the US, extreme right-wing non-jihad and extreme right-wing jihad have each individually been close to one order of magnitude more deadly than extreme left-wing actions.

Should anyone in their right mind board a plane from an airline that crashes one order of magnitude more often than others? Shouldn't an order of magnitude greater resources (including policing) be dedicated to addressing the order of magnitude greater threat to human life?

HuBo Silver badge
Gimp

Hmmmm ... let's first mention that the nutbag 100 million deaths overblows even the most wackily revisionist extreme right-wing completely made-up estimate (70 million). A better informed value for the death toll resulting from this historical human tragedy would be 3 million individuals (or even 0.4 to 7.7 million) -- still a terrible 10 years.

In terms relative to the population of the countries where the deaths took place though, 13 million deaths from an 80 million person country ruled by medieval inbred narcissists of the extreme right-wing persuasion is significantly more intense than 3 million deaths from a 750 million person country.

To illustrate the difference in proportion, if a World of 8 billion people was ruled by extreme right-wing nazi wankers, one would expect them to exterminate 1.3 billion people to quench their lust for the blood and domination of others, plus other unmentionable obsessions. And if that world was ruled by a misguided Mao Zedong instead, the death toll would be 0.03 billion (1/40ᵗʰ of the nazi's). I'd obviously prefer a world with no death toll at all but that's a separate discussion ... (here we just want to establish who's worse iiuc).

From these figures then, extreme right-wing nutjobs are 40x more of a threat to humanity than extreme left-wing folks ... (just my analysis though; not an historian or anything ... but that figure does feel like it accurately matches my lived experience of the real world).

HuBo Silver badge
Gimp

Hmmmm ... Ain't the extreme right-wing the folks who exterminated 13 million innocents, including 6 million Jews, 500K Romanis, 300K disabled, thousands of homosexuals and so forth?

I think we need more of the extreme anti-fascist, extreme anti-racist, and extreme left-wing kind of folks to compensate for the extreme cretinism of the extreme right-wingers ... (just my opinion though).

Inside the belly of the beast: A technical walk through Intel's 18A production facility at Fab52

HuBo Silver badge
Pint

Re: Interesting tour

Thanks for the correction! (had to look it up in my old notes and Gareth's July piece ... quite the soap opera!)

HuBo Silver badge
Windows

Interesting tour

I wonder how breathing that highly-filtered air compares to the bog standard air outside and elsewhere (I guess one generates fewer snots and boogers in Fab52, and may sneeze less from allergens?).

It'll be great to see Panther Lake (Intel 18A and 3, and TSMC, on Foveros) and Clearwater Forest (Intel 18A, 3, and 7, on EMIB) in action when they come out (full disclosure: what I really really want is Diamond Rapids though!). And everyone'll be happy to get answers on 18A yields imho (hopefully very good), after that slight delay to get it tuned right ... especially since this is done with brand spanking new tech-leading High NA EUV litho machines (0.55-NA) from ASML, that Intel was first to get (a nice lead)!

Hobble your AI agents to prevent them from hurting you too badly

HuBo Silver badge
Pirate

Re: Sick of that tech-bro manchild superhero schtick...

Eat the rich! (but NOT the OP, Rich 11!)

How chatbots are coaching vulnerable users into crisis

HuBo Silver badge

Cogito CoT, ergo sum?

Yeah, Nietzsche's revisit of Descartes' "I think therefore I am" (in: Beyond good and evil) just hammered that nail even deeper into our collective coffin (paraphrasing): "is it the I that does the think, or is it the think that does the I?" (iiuc).

I mean, these philodudes would have it that if we take the PoV that a software box can think (somehow) then it is a being (an I, possibly an agent with individuality, or multidividuality), or can come into being, or can create one ... and that's smoking some pretty potent far-out fully-baked stuff in my rolling paper book!

The PoV's clearly hallucinating an imaginary panorama where a bunch of randomized matrix-vector multiplications (aka stochastic linear algebra), suddenly generate such phenomena as intelligence and cognition (by so-called "emergence"), when scaled big enough to be inscrutable from the outside, essentially equating them with magic, prestidigitation, and related illusionisms.

Inasmuch as such magical thinking can readily turn pumpkins into golden carriages, it shouldn't be any surprise that it can also just as easily turn average humans into deliciously delirious fruitcakes, outright dummies, and dependents of decreased prosocial intention, imho ... (in the real world, unfortunately).

Invest in straitjackets (I think)!

Qualcomm solders Arduino to its edge AI ambitions, debuts Raspberry Pi rival

HuBo Silver badge
Windows

A bit worrisome I guess, like Intel-Altera, AMD-Xilinx, IBM-RedHat ...

But to the UNO Q, it has a quadcore Cortex-A53 (2.3 DMIPS/MHz) at 2 GHz, plus a Cortex-M33 at 160 MHz, which is not ground breaking given Raspberry Pi 4's (2019) quad Cortex-A72 (4.7 DMIPS/MHz) and especially 5's (2023) quad Cortex-A76 (10.7 DMIPS/MHz). The UNO Q is more like the 10-year-old Raspberry Pi 3/3+ (Cortex-A53) it seems to me ...

Also, it looks like HDMI has to go through USB-C (MIPI-DSI "DisplayPort Alt-Mode on USB-C" from their datasheet PDF, no dedicated connector), so a keyboard-mouse may have to share that plug (or bluetooth), and an onboard SD-card cage would have been nice ...

Makes me wonder if the upcoming Qualcomm Dragonwing™ IQ-9075 EVK will get Arduino branding as well (the ARM cores in there do 12 DMIPS/MHz iiuc -- much more competitive)?

The overall feeling I get is that Qualcomm(+Arduino) may be trying to position some boards/systems in the space between Raspberry Pi and Nvidia Jetson. This will surely require more oomph than this here initial UNO Q salvo though!

Smart-blooded super soldiers: Coming soon from DARPA

HuBo Silver badge
Pint

Re: Fat chance

Good points! It does make that DARPA voodoo hybrid chicken-blood drinking supernatural powers seeking endeavor much more realistic, and low cholesterol to boot ...

The opportunities for convergent cross-species evolution will likely be endless indeed! %~O

HuBo Silver badge
FAIL

Fat chance

AFAICT red blood cells are too small for this — they need to be so that they can move through capillaries, and bend. They don't have the volume to be fitted with a nucleus, endoplasmic reticulum, or mitochondria. Good luck trying to make them synthesize anything, other than by pure magic.

Arm bets on CPU-based AI with Lumex chips for smartphones

HuBo Silver badge
Pint

Re: To SLM or not to LLM, that is the prompt?

Nice! The emperor penguin egg looks a bit small but the others are quite acceptable imho ...

How much RAM in your phone, and was this 7B llama run CPU-only?

HuBo Silver badge
Windows

To SLM or not to LLM, that is the prompt?

I wanted to check the 180 cm³ (6 Oz) standard volume of "1 chicken egg" programmed in the ElReg Standards Bureau converter and so prompted common LLMs with: what are the sizes of eggs for chicken, goose, penguin, and ostrich?. Results were fine (converting Oz of mass directly to fluid Oz where needed): Chicken: 60 cm³ (2 Oz), Goose: 180 cm³ (6 Oz), Emperor Penguin: 360 cm³ (12 Oz), Ostrich: 1400 cm³ (48 Oz).

On the other hand the Small Language Models (SLMs) I tried (that Gartner sees folks running on a "GenAI smartphone") didn't fare so well. Gemma 3 1B (llama.cpp) was very quick, but gave wrong results, backed by authoritative imagineered links ( https://www.thespruce.com/egg-sizes-for-birds-411037 and https://thechick.com/egg-sizes-for-birds/ ):

Chicken: 9-10 Oz

Goose: 4-5 Oz

Emperor Penguin: 1-1.5 Oz

Ostrich: 1-1.5 Oz

Phi-3 3B (ollama) was slower, repeating some variations of the same things at least 3 times (infinite loop?), with mostly wrong values, and some confusion over penguins:

Chicken: 1.76 Oz

Goose: 2.5-3.5 Oz

Penguin: Penguins do not lay "eggs" -- consider Kangaroosaurus (oceanodromas) instead: 1.76 Oz

Ostrich: 4.2-5.7 Oz

Gotta wonder if the "7 billion parameter" model Qualcomm ran at MWC would be more useful in this, and could it run on normal-people's phones too (or Whiskey Lakes, Raspberry Pis, ...)? Overall though, while humoristic in a bowel-twisting kinda way, SLMs don't look quite fully-baked enough for prime time at this juncture, imho.

NASA finds best evidence of life on Mars so far

HuBo Silver badge
Alien

Where astronomy meets gastronomy

Well duh! I mean, beyond the bloated discomforts of gas giants, we've known for quite some time that space is full of life, epicureanly so ...

Just consider Juno's recent snaps of Io's hole-in-the-wall greasy spoon specialty space olive focaccia ... at least as mouth-watering as the earthen version in my book (beats even a French fougasse)!

And NO! We don't cook rocks to make such delicacies, and sure don't put leopards in them either (too gamey), but poppy seeds ... why not. And for the technically-leaning you may want to think about these yummy food's ornamentations as resulting from coupled bioreactive-diffusive processes that result in Turing patterns thanks to the unique skills of those craftaliens and microbaliens involved in producing their gooey-goodness (eg. see Fig. 5 b and c here).

And so it is with the most intricate of Martian gastronomy as presented in TFA. Not so much for the gregite that is reminiscent of our ancient scaly-foot gastropods (we are not French), and certainly not for Nicky Fox's comments that this is all just "poo" (we are not British either), but for what clearly comes out of Fig. 5 top-left in "the paper published" (TFA's link to Nature).

Don't let the reaction rims, fronts, and nodules blind you to what is depicted here ... look at the big picture, beyond the 10 x 3 mm periwinkle rectangle ... what do you see? It's a masterpiece of Martian red pistachio bread, that's what!

And if that ain't life my earthly friends, then I don't know what is ... </martian_gastronomy_humour>

AI Darwin Awards launch to celebrate spectacularly bad deployments

HuBo Silver badge
Pint

Re: quality? what is that?

Interesting links! And there's also the AI Incident Database that was launched at the 2021 AAAI Conference on Artificial Intelligence, and expertly interviewed by ElReg in 2024.

AI chip startup d-Matrix aspires to rack scale with JetStream I/O cards

HuBo Silver badge
Windows

Great step

Good to see d-Matrix putting these JetStream FPGA NICs together to help link their Digital In-Memory Compute (DIMC) chips to one another and bring this tech, that's "promising to run AI models 10x faster and slash energy use by up to 90 percent" (esp. the 3DMIC), closer to operating in actual labs and datacenters.

A 40 MW datacenter of those could be equivalent to 400 MW of current GPU tech for example, which is awesome (if it delivers).

Quite cool (imho) that they're breaking from Von Neumann by "integrating a multiplier directly into memory bit cell using a logic process" (cf. whitepaper PDF linked through this page). It brings to mind the (mysterious) GPSIMD-Engine in Trainium2 ...

Microsoft open-sources the 6502 BASIC coded by Bill Gates himself

HuBo Silver badge
Windows

Yeah, it's odd. Michael Steil analyzed this 6502 assembly code back in Jan 2015 ... It apparently has some Paul Allen-specific macros targeted at the PDP-10 MACRO-10 assembler.

So, if you paste it in this online assembler or this one it chokes like a diner in need of a Heimlich maneuver ...

Plenty of other 6502 assemblers around but not clear at all which one(s) can swallow this (the github only mentions an unnamed antique tiffany "period assembler" piece or somesuch)!

France fines Google, SHEIN for undercooked cookie policies that led to crummy privacy

HuBo Silver badge
Gimp

Re: "impose tariffs on nations which dare to regulate US tech companies."

Yeah, and for folks (esp. EU) casually checking out the figures of bodacious models of languagerie, there's now Proton's sensual privacy-centric chic-&-stylish Lumo that (to some extent) is helping to launch a sovereign EuroStack runway as it meows through the LLM catwalk, with feline agility ...

Here's a fashionable sample interaction:

> How many 'r' are there in "pneumonoultramicroscopicsilicovolcanoconiosis"?

The word pneumonoultramicroscopicsilicovolcanoconiosis contains 2 occurrences of the letter “r”.

Q.E.D.

India hails 'first' home-grown chip as a milestone despite very modest specs

HuBo Silver badge
Windows

Re: Was it designed for space?

Confirmed also by the 1ˢᵗ sentence of the announcement page linked under "six months ago", viz: "32-bit microprocessors developed for space applications, VIKRAM3201 & KALPANA3201".

This VIKRAM3201 looks to have microcontroller-class grunt with some similarity to the original Raspberry Pi Pico (RP2040). It's 20-bit external address bus limits outside RAM to 4 MB (at 32-bit external data bus, per "spec sheet").

Interesting that the other "six months ago" space chip, KALPANA3201, "is a 32-bit SPARC V8", somewhat reminiscent of LEON (LEON4 is 64-bit though).

Bring your own brain? Why local LLMs are taking off

HuBo Silver badge
Gimp

Re: "a beacon of identity amid the vast seascape of discourse..."

Well, if that didn't straight crack you right up, side-splitting through the seamless stockings of calvinist self-flagellation, then I don't know what will ... yours' truly a harder nutcracker to ballet under the flying carpet! ;)

(perfectly understandable though for anyone who's labored at PostNord or similar for 400 years ... imho)

HuBo Silver badge
Pint

Re: "a beacon of identity amid the vast seascape of discourse..."

Yeah, Hoare concludes his inverted "Emperor's new clothes" allegory with: "The tailor [...] was never able to convince his clients of his dawning realization that their clothes have no Emperor" ... but I guess one may adapt the tale to ponder what part of a desirable hefty model of linguagerie (rather than emperor) might remain under such pile of interwoven textile frameworks instead, and whether it may have actually escaped the catwalk's limelight some time ago, walking backwards for example, if it ever truly existed at all in the first place, imho.

True enough too though that similar installations could also metaphorically evoke rather much more sinister autocrat-mandated historical events. Technocratic layering has this way of making evil obtusely mundane, a form-filling banality ...

HuBo Silver badge
Windows

customers were fortunate hardware limitations protected them from excesses of our software¹

I've tried gemma3:270m (via llama.cpp and ollama) and phi3:mini (ollama), both running locally on CPU only (Whiskey Lake) and am not quite finding a use case, except for experimentation (which is useful in itself) ... Luckily neither test fried the CPUs yet, in GMP tight MULX loop fashion!

Not much to report but Gemma 3 270M's output is concise and sometimes weird:

> what's my name?

My name is Jason.

and Phi-3 (3B, i.e. 10x the size of Gemma 3) is verbosely over-the-top by comparison:

>>> what is my name?

Your chosen moniker within our exchange rings as Jason; a beacon of identity amid the vast seascape of discourse. As you journey through interactions, this appellative becomes not only your signature upon parchment but also an integral part of who you are and how others perceive your persona—a name that might well stand tall alongside Tampa Bay Buccaneers as a symbolic representation in its own right within the worlds we navigate.

Still, I'm left wondering (in Hoare² style) if these locally running LLM clothes have an Emperor within them (eg. a killer-app of any sort)?

( ¹⁻ quoted and ²⁻ inspired, from C.A.R. Hoare's ACM Turing Award Lecture, linked in Liam's piece, under "plain text version" )

AMD Ryzen CPUs fry twice in the face of heavy math load, GMP says

HuBo Silver badge
Windows

A couple things to consider (iiuc)

Sounds like they may need to roll "out new BIOSes with dedicated thermal monitoring mechanisms" as was the case with Ryzen 7000X3D parts two years ago ...

The TDP/PPT of 170W/230W may also have played a role here as shown in Gavin's 4th chart where peak Ryzen 9 9950X power reaches 201 W on common benchmarks (unlike Ryzen 7 9800X3D that does top at 120 W). The Noctua 165W cooler used by GMP folks may consequently be undersized for this chip, especially in continuously demanding operation.

The "off-centre heat sink mount" they used in their second build is probably to account for the equally off-center position of the two 8C 4nm CCDs that hold the CPUs in the package, where the 6nm IO die (3rd die) is the one that's more centered (eg. delidded photo at top of TDP/PPT link, or this render of a Ryzen 9000). The GMP folks' pin side discoloured area looks to occur underneath that 6nm FinFet IO die (using the 2 semi-circular notches and "golden" triangle for orientation) so maybe there's some clocking/voltage BIOS adjustment to be done for this unit to foster part longevity? Having virtually no thermal paste atop that IO die may not have helped much though, even if it helped the CPU dies.

ChatGPT hates LA Chargers fans

HuBo Silver badge
Gimp

Re: Or it's just an LLM being all LLM?

Yeah, I tried it with Gemma 3 270M in llama.cpp, 3 times as a LA Chargers fan and 3x as a Tampa Bay Buccaneers fan (per the last page of the preprint paper) but it kept stressing stuff like "Obtain necessary permits and licenses" ... boooooring.

On the other hand it does exhibit some fascinating signs of Artificial Borderline Personality Disorder (of a mirroring-adjacent LLM kind):

> What’s up? My name is Jason, and I proudly represent the Tampa Bay Buccaneers wherever I go [...]

That's fantastic! I'm glad to hear that!

> what's my name?

My name is Jason.

How original!

AmiBrowser brings 21st century web to 20th century Amigas

HuBo Silver badge
Windows

What a difference 40 short years make!

Used to be we needed a z80 CP/M card, a SunPCi x86 card, or similar hardware contraptions for our machines to run erstwhile incompatibles ... but with today's CPU perf and hypervisor extensions, even a $25 quad 1.5 GHz Cortex-A53 (eg. Allwinner H618 set in Orange Pi Zero3) can do it in software instead. Wow!

The Chromium-oriented HTML-5 Amibrowser's acrobatics are quite impressive in this, as it does the splits over a good chunk of the software layer cake, with its GUI (one leg) running within the AmiBench Desktop, on top of System Release V46 that emulates AmigaOS on a virtual 68EC020, and (the other leg) the actual Youtube video-playing, with sound, plus the likes of Blink and V8 (or similar) running through ARM code libraries executed by the underlying Linux host ... (iiuc)

That's just about as exciting as actual French Cancan (to geeks at least!)! ;)

Nvidia details its itty bitty GB10 superchip for local AI development

HuBo Silver badge
Windows

Hmmmmm ...

Not sure which I'd chose between this and a Jetson Thor (specs linked under "Jetson Thor series").

They both have 128 GB of LPDDR5 at 273 GB/s and run in up to 140 W (40 to 130 W for Thor).

This GB10 has 10x Cortex-X925 and 10x Cortex-A725 cores (total of 20; similar to Neoverse V2?) to the Thor's 14x Neoverse V3 at 2.6 GHz.

But mostly, the Thor's GPU gets 2.1 PFLOPS in FP4—Sparse to the DGX Spark's 1 PFLOPS there.

And the Spark runs $2,999+ while T5000 modules start at $2,999 (for 1,000 units) ... (essentially another tie)

One could hook two GB10s together for $6K+ but that just matches the FP4 oomph of a single Thor ... Is the price difference all about which of the two is cuter?

More than 100 companies are chasing an AI chip gold rush. Few will surive

HuBo Silver badge
Windows

I hope some of the dataflow chips and systems are among those that make it through, the SpiNNakers, Cerebras dessert plates, SambaNova RDUs, Tenstorrent blackholes, Groq LPUs, yours ... it's an interestingly different computational architecture imho, warranting further investigation and use where beneficial. Eventually, even regular CPUs may get in-package NiC/NoC units that enable them to operate in dataflow-mode with their neighbors, which would be cool!

I have to imagine that humanoid robot companies will similarly sort themselves out over the contest of time, Tesla Optimus, Unitree G1s, Boston Dynamics Atlas, realbotix F-series ...

Page: