Jump to content

Neural Engine on M1 is amazing...

TheReal1980

We have read all about how good the M1 is in all sorts of ways but something not spoken about as much is the M1's Neural Engine, with 16 Cores and the capability of doing 11 trillion operations per second.

We are now seeing what can be done in the picture software Pixelmator as an example.

 

 

 

My thoughts

To me it seems the M1 is an even bigger deal than most people understand. To be able to pack that kind of punch in a fanless laptop is insane and this Neural Engine is just the icing on the cake.

I wonder how Intel, AMD and different mobile chip developers are thinking right now.

 

 

Sources

https://www.pixelmator.com/blog/2019/12/17/all-about-the-new-ml-super-resolution-feature-in-pixelmator-pro/

If it ain´t broke don't try to break it.

Link to comment
Share on other sites

Link to post
Share on other sites

I've not really looked at the ML Cores on my M1 MacBook Air but this is easily an amazing feature if supported by the software manufacturers! 

Main Machine:  16 inch MacBook Pro (2021), Apple M1 Pro (10 CPU, 16 GPU Core), 512GB SDD, 16GB RAM

Gaming Machine:  Acer Nitro 5, Core i7 10750H, RTX 3060 (L) 6GB, 1TB SSD (Boot), 2TB SSD (Storage), 32GB DDR4 RAM

Other Tech: iPhone 15 Pro Max, Series 6 Apple Watch (LTE), AirPods Max, PS4, Nintendo Switch, PS3, Xbox 360

Network Gear:  TP Link Gigabit 24 Port Switch, TP-Link Deco M4 Mesh Wi-Fi, M1 MacMini File & Media Server with 8TB of RAID 1 Storage

Link to comment
Share on other sites

Link to post
Share on other sites

Cool, but RTX GeForce had Tensor Cores for year now and AMD have them own version on them up coming GPU. Also this type of system have been in android phone for some time like in the Huawei Mate 40 Pro.  It other words Android and PC been having this type of system for years and apple just catching up.

 

There nothing new. 

Link to comment
Share on other sites

Link to post
Share on other sites

14 minutes ago, TheReal1980 said:

11 trillion operations per second

Boasting about numbers yet doesn't compare it to anything.

A PC Enthusiast since 2011
AMD Ryzen 7 5700X@4.65GHz | GIGABYTE GTX 1660 GAMING OC @ Core 2085MHz Memory 5000MHz
Cinebench R23: 15669cb | Unigine Superposition 1080p Extreme: 3566
Link to comment
Share on other sites

Link to post
Share on other sites

3 minutes ago, A51UK said:

Cool, but RTX GeForce had Tensor Cores for year now and AMD have them own version on them up coming GPU. Also this type of system have been in android phone for some time like in the Huawei Mate 40 Pro.  It other words Android and PC been having this type of system for years and apple just catching up.

 

There nothing new. 

The A11 chip released for iOS devices in 2017 had fixed function hardware for accelerating machine learning workloads (INT8/FP16 matrices), at this time nvidia hadn't even released the Volta architecture. This is simply misinformation.

Platform agnostic software engineer & small business owner. 

Link to comment
Share on other sites

Link to post
Share on other sites

3 minutes ago, A51UK said:

Cool, but RTX GeForce had Tensor Cores for year now and AMD have them own version on them up coming GPU. Also this type of system have been in android phone for some time like in the Huawei Mate 40 Pro.  It other words Android and PC been having this type of system for years and apple just catching up.

 

There nothing new. 

Apple started shipping the NE in every new iphone/ipad starting with the A11 Bionic in 2017.

Now in every new Mac.

Completely different chances of app developers leveraging it in everyday usage for regular people.

But go ahead with your youtube-comment style superficial responses. “Catching up”..

Link to comment
Share on other sites

Link to post
Share on other sites

4 minutes ago, saltycaramel said:

Now in every new Mac.

Completely different chances of app developers leveraging it in everyday usage for regular people.

 

That’s what some people don’t get about these AS Macs.

 

Imagine a software ecosystem where that CPU power is the baseline, a given.

Imagine a software ecosystem where that iGPU power is the baseline, a given.

Imagine a software ecosystem where the presence and the speed of that Neural Engine is the baseline, a given.

Imagine a software ecosystem where the speed of that Unified Memory is the baseline, a given.

Imagine a software ecosystem that can assume that the slowest internal storage to EVER ship was 2500MB/s. 

 

What will software leveraging all of this look like in a few years?

What this being the baseline, the bottom of the barrel, the lowest end will imply for developers designing an app?

Link to comment
Share on other sites

Link to post
Share on other sites

14 minutes ago, A51UK said:

Cool, but RTX GeForce had Tensor Cores for year now and AMD have them own version on them up coming GPU. Also this type of system have been in android phone for some time like in the Huawei Mate 40 Pro.  It other words Android and PC been having this type of system for years and apple just catching up.

 

There nothing new. 

You're being very disingenuous right here.

  • On PCs, we have only had NPUs for 1 generation of GPUs, and only from Nvidia. AMD and Intel do not have NPUs in any currently available consumer products. Much less in actual products like laptops.
  • If we are going to bring mobile phones into the discussion like you did then this is not the first time Apple uses NPUs either.
  • We have "had this type of system for years" (from one manufacturer) and we barely see it used in any applications. Meanwhile, Apple launched their computer with it yesterday and multiple apps are already using it.
  • Just because something isn't new doesn't mean it is bad. The 3070 and RX 6800 for example do not really offer anything "new" either. It's performance that was possible to get several years ago, and mostly features that were available several years ago too. But despite that, it is news when it gets announced and released. Right?
Link to comment
Share on other sites

Link to post
Share on other sites

As I wrote in some other thread surrounding the M1 announcement.

It will be interesting to see how developers will leverage the power of the NE that is built in to the base system in other applications. It should be able to provide a compute boost even in non ML applications as long as you write code to use it. 

Link to comment
Share on other sites

Link to post
Share on other sites

32 minutes ago, A51UK said:

Cool, but RTX GeForce had Tensor Cores for year now and AMD have them own version on them up coming GPU. Also this type of system have been in android phone for some time like in the Huawei Mate 40 Pro.  It other words Android and PC been having this type of system for years and apple just catching up.

 

There nothing new. 

Think ASR and TTS. Just to get to "Siri" level of functionality still requires a high end GPU with a lot of RAM. The stuff that runs on laptops and phones is entirely in the cloud. Just to run a neural TTS system requires 4GB of video memory with a single voice, alone. You don't necessarily need all the cores if you're not training a system, but you still need some to use it, and the CPU is phenomenally bad at it, even with all the AVX instructions.

 

Likewise for OpenCV/ARKit, requires some level of functionality here because running it entirely on the CPU is going to have poor performance.

 

To that end, I think this is something that needs to be looked at vs comparable nvidia, amd and intel solutions and since there is no common middleware to test that, it might not be be doable for some time.

Link to comment
Share on other sites

Link to post
Share on other sites

Benvista PhotoZoom didn't use ML and it had spectacular results years ago before ML was even a thing anywhere. It made some insane results with upscaling. If Bicubic or Lanczos looked blocky and blurry, PhotoZoom's S-Spline Max delivered massive detailed and sharp images that looked light years better than anything else. Worth chekcing out.

Where ML might shine is computational time though...

Link to comment
Share on other sites

Link to post
Share on other sites

Interesting.. Pixelmator specifically was called out in MKBHD's video for not working on his M1 mbp. Guess an update happened between him trying it and now.

🖥️ Motherboard: MSI A320M PRO-VH PLUS  ** Processor: AMD Ryzen 2600 3.4 GHz ** Video Card: Nvidia GeForce 1070 TI 8GB Zotac 1070ti 🖥️
🖥️ Memory: 32GB DDR4 2400  ** Power Supply: 650 Watts Power Supply Thermaltake +80 Bronze Thermaltake PSU 🖥️

🍎 2012 iMac i7 27";  2007 MBP 2.2 GHZ; Power Mac G5 Dual 2GHZ; B&W G3; Quadra 650; Mac SE 🍎

🍎 iPad Air2; iPhone SE 2020; iPhone 5s; AppleTV 4k 🍎

Link to comment
Share on other sites

Link to post
Share on other sites

7 minutes ago, Video Beagle said:

Interesting.. Pixelmator specifically was called out in MKBHD's video for not working on his M1 mbp. Guess an update happened between him trying it and now.

Yeah they came out with an update afterwards.

If it ain´t broke don't try to break it.

Link to comment
Share on other sites

Link to post
Share on other sites

56 minutes ago, saltycaramel said:

What this being the baseline, the bottom of the barrel, the lowest end will imply for developers designing an app?

Probably the same as 2 decades ago, a bunch of quickly thrown together and unoptimized software that runs like shit, worse than good software running on hardware that doesn't have all those bells and whistles...

That's what tends to happen when "we've got more power than we need, let's use it to save us dev effort"

F@H
Desktop: i9-13900K, ASUS Z790-E, 64GB DDR5-6000 CL36, RTX3080, 2TB MP600 Pro XT, 2TB SX8200Pro, 2x16TB Ironwolf RAID0, Corsair HX1200, Antec Vortex 360 AIO, Thermaltake Versa H25 TG, Samsung 4K curved 49" TV, 23" secondary, Mountain Everest Max

Mobile SFF rig: i9-9900K, Noctua NH-L9i, Asrock Z390 Phantom ITX-AC, 32GB, GTX1070, 2x1TB SX8200Pro RAID0, 2x5TB 2.5" HDD RAID0, Athena 500W Flex (Noctua fan), Custom 4.7l 3D printed case

 

Asus Zenbook UM325UA, Ryzen 7 5700u, 16GB, 1TB, OLED

 

GPD Win 2

Link to comment
Share on other sites

Link to post
Share on other sites

16 minutes ago, Kilrah said:

Probably the same as 2 decades ago, a bunch of quickly thrown together and unoptimized software that runs like shit, worse than good software running on hardware that doesn't have all those bells and whistles...

That's what tends to happen when "we've got more power than we need, let's use it to save us dev effort"

Yeah, just like iOS/iPadOS apps compared to Android apps, am I rite? 🥱

Link to comment
Share on other sites

Link to post
Share on other sites

Also: if everything moved to the web or cross-platform desktop webapps, Apple CPUs would still be preferable being the absolute champions at crushing every web/html/js benchmark..

Link to comment
Share on other sites

Link to post
Share on other sites

17 minutes ago, saltycaramel said:

Also: if everything moved to the web or cross-platform desktop webapps, Apple CPUs would still be preferable being the absolute champions at crushing every web/html/js benchmark..

And battery life

Exciting times xD

Link to comment
Share on other sites

Link to post
Share on other sites

2 hours ago, LAwLz said:

Intel do not have NPUs in any currently available consumer products.

Playing devil's advocate, Intel has had the movidius for quite a while now, and Google has the Coral as a competitor. 

 

However, those are pretty useless for the general consumer since most software won't make use of it and it's more of a dev/embedded thing 😜

FX6300 @ 4.2GHz | Gigabyte GA-78LMT-USB3 R2 | Hyper 212x | 3x 8GB + 1x 4GB @ 1600MHz | Gigabyte 2060 Super | Corsair CX650M | LG 43UK6520PSA
ASUS X550LN | i5 4210u | 12GB
Lenovo N23 Yoga

Link to comment
Share on other sites

Link to post
Share on other sites

4 hours ago, saltycaramel said:

 

That’s what some people don’t get about these AS Macs.

 

Imagine a software ecosystem where that CPU power is the baseline, a given.

Imagine a software ecosystem where that iGPU power is the baseline, a given.

Imagine a software ecosystem where the presence and the speed of that Neural Engine is the baseline, a given.

Imagine a software ecosystem where the speed of that Unified Memory is the baseline, a given.

Imagine a software ecosystem that can assume that the slowest internal storage to EVER ship was 2500MB/s. 

 

What will software leveraging all of this look like in a few years?

What this being the baseline, the bottom of the barrel, the lowest end will imply for developers designing an app?

IB4 more Javascript shenanigans and unoptimization... 

Link to comment
Share on other sites

Link to post
Share on other sites

The real challenge is going to be driving adoption of the CoreML library to benefit from this inclusion. Lots of data science/ML work is done with Python, but as far as I know there are no python libraries that have been updated to support the hardware acceleration yet. Even some other very crucial libraries like TensorFlow do not currently have support (in fact, TensorFlow on macOS doesn't support GPU acceleration because Metal doesn't support templetized kernels). 

 

Bit of a road ahead, but still promising. 

15" MBP TB

AMD 5800X | Gigabyte Aorus Master | EVGA 2060 KO Ultra | Define 7 || Blade Server: Intel 3570k | GD65 | Corsair C70 | 13TB

Link to comment
Share on other sites

Link to post
Share on other sites

Apple's custom hardware/software integration in terms is so far ahead of everything else in the market. Some might be able to leverage custom SoCs like Samsung's exynos line, some might be able to hyper-focus on software, like Google, but nobody can match Apple's silicon-software fusion and this is another example of that. A fantastic neural engine on the hardware side coupled with what is looking like a good software base that can only get better.

Link to comment
Share on other sites

Link to post
Share on other sites

5 hours ago, StDragon said:

A "Deepfake" accelerator.

 

2021 is going to be an interesting year.

Instagram fake influencers are already going wild with all the sharp jaw lines they could have and big eyes and it'll all be perfect lol

Link to comment
Share on other sites

Link to post
Share on other sites

1 hour ago, RejZoR said:

Instagram fake influencers are already going wild with all the sharp jaw lines they could have and big eyes and it'll all be perfect lol

how big?

✨FNIGE✨

Link to comment
Share on other sites

Link to post
Share on other sites

14 hours ago, saltycaramel said:

magine a software ecosystem where that CPU power is the baseline, a given.

Imagine a software ecosystem where that iGPU power is the baseline, a given.

Imagine a software ecosystem where the presence and the speed of that Neural Engine is the baseline, a given.

Imagine a software ecosystem where the speed of that Unified Memory is the baseline, a given.

Imagine a software ecosystem that can assume that the slowest internal storage to EVER ship was 2500MB/s. 

this is all hardware.

 

Imagine a hardware ecosystem where 16gb memory is the max.

AMD blackout rig

 

cpu: ryzen 5 3600 @4.4ghz @1.35v

gpu: rx5700xt 2200mhz

ram: vengeance lpx c15 3200mhz

mobo: gigabyte b550 auros pro 

psu: cooler master mwe 650w

case: masterbox mbx520

fans:Noctua industrial 3000rpm x6

 

 

Link to comment
Share on other sites

Link to post
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now

×