The Hard Tradeoffs of Edge AI Hardware

  Рет қаралды 87,317

Asianometry

Asianometry

Күн бұрын

Errata:
I said in this video that "CPUs and GPUs are not seen as acceptable hardware choices for edge AI solutions". This is not true, as CPUs are commonly used for small, sub-$100 items. And GPUs are frequently used in lieu of FPGAs due to their ease of programming. Thanks to Patron Gavin for his input.
Links:
- The Asianometry Newsletter: asianometry.com
- Patreon: / asianometry
- The Podcast: anchor.fm/asianometry
- Twitter: / asianometry

Пікірлер: 210
@philippepanayotov9632
@philippepanayotov9632 Жыл бұрын
Yet another interesting video, which condenses a lot of information into a manageable chunk. Keep it up!
@tempacc9589
@tempacc9589 Жыл бұрын
As someone who actually deploys edge AI I heavily disagree with how "easy" you make FPGA's and ASICS seem. For the vast majority of "smaller" projects the Nvidia Jetson series is a far better choice since they support newer algorithms and functions. Especially with the speed the field of AI is progressing at. Furthermore fp16 GPU tensor cores are basically optimized for ML inference and provide good performance if you want to spend a little extra time on converting the model, though often even that has compiling issues with newer ML models.
@g.l.5072
@g.l.5072 Жыл бұрын
Agreed.
@adissentingopinion848
@adissentingopinion848 11 ай бұрын
As a verification engineer for FPGAs, let me tell you that the Hardware Description Languages you need to use are ancient relics based off of PASCAL. They have so many pitfalls and inexplicable limitations that cause more problems than the design itself. And they need to be so complex because the design space for this stuff is extreme. When the compiler(synthesis+PAR) has to conjure up a complete mapping of billions of custom registers and their connections, you're facing the limitations of hardware, physics, software, development tools, and whatever hellish design process your company is in. If you need an FPGA for a problem, you need to be damn sure there isn't a specific ASIC or a microprocessor that can do it faster or cheaper for you. They are good for super fast timing, fast turnaround for design iterations, a pathway to ASIC dev, and as much parallelization as you can stuff through an interface (the main bottleneck!). And if you want these benefits, be prepared to build it all from the ground up. Generalized cores are liable to need tuning for extra speed, and if you're using an FPGA you're gonna need to milk it for every last drop or downgrade to a cheaper chip because cost will be king.
@parkerbond9400
@parkerbond9400 2 күн бұрын
Asianometry makes EUV lithography understandable, when we all know it's magic. Making very hard problems seem easy is just par for the course.
@petermuller608
@petermuller608 Жыл бұрын
Dude your content is as outstanding as it is random. From soviet oil export over chip manufacturing to edge AI models xX
@fennewald5230
@fennewald5230 Жыл бұрын
I write inference pipelines on jetsons for my work. The latest generation have some very attractive performance characteristics. They do pull 60W, which definitely isn’t nothing, but for our use case it’s manageable. Something that wasn't stressed is that those devices are heavily optimized around fixed point 8 bit throughput. A high-spec Orin can put out 275 TOPS, which is more than triple a 4090s 83 TFLOPS. Even if the models are much weaker, the increase in throughput opens up a lot of flexibility with system design.
@azertyQ
@azertyQ Жыл бұрын
Qualcomm AI100's can do 400TOPS @ 75W
@Smytjf11
@Smytjf11 Жыл бұрын
@@azertyQ yeah, but an Orin costs $2K
@StephenGillie
@StephenGillie Жыл бұрын
Post-training pruning is very much how human neural networks learn - massive connections anywhere and everywhere over the first few years - during the initial training phase, then massive pruning to get rid of the unnecessary connections.
@jadeaffenjaeger6361
@jadeaffenjaeger6361 Жыл бұрын
This is true and underlines the vast differences between our brains and machine learning hardware: Unstructured Pruning for Neural Networks tends to struggle with performance because it usually maps poorly to the existing highly-parallel and highly structured hardware. In the industry, methods like low-bitwidth quantization and custom architectures with low parameter counts (e.g. MobileNet as mentioned in the video) tend to see more use because their regular structure can exploit parallel hardware a lot better.
@TheDiner50
@TheDiner50 Жыл бұрын
It is not even years. As soon as one got enough training for the basic task the pruning beings quickly until your ok at it. You are at a baseline training level. It is first months or years of small improvements stacking up that starts to really be noticeable. Much harder but a sign of real taut skills from doing it for real. Usually people that are really good at something look at you funny with a long stair into the distance when you comment on there 'skill' and effortless execution. Since the improvements are many times just work to them and really rather not do it in the first place. The only way a AI can be smart is if it learns to do stuff on it's own faster or better in some way without the help of the creator going in messing with it constantly.
@MyPhone-qg2eh
@MyPhone-qg2eh 6 ай бұрын
Only the relevant solutions need to be deployed. And can be accessed by a simple sorting algorithm most of the time.
@TheTyno37
@TheTyno37 Жыл бұрын
My fave video of yours recently! Thanks for making it!
@nekomakhea9440
@nekomakhea9440 Жыл бұрын
for a second, I thought the cow photo at the beginning said "dont eat ass"
@MarcdeVinck
@MarcdeVinck Жыл бұрын
That solder bridging in the stock video at 2:56 and 3:00. Yikes! Great video BTW!
@Andrew-pc8zn
@Andrew-pc8zn Жыл бұрын
Wake up babe! Asianometry posted!
@frazuppi4897
@frazuppi4897 Жыл бұрын
Amazing video (as always), I am a ML engineer, it's not (always) true that the less weight the model has the less memory/faster it is. I know it's a little counter intuitive. It all comes down to how many GFLOPs the model need on some hardware
@davidcarlsson1396
@davidcarlsson1396 Жыл бұрын
I made a thesis at university for graduation, basically image processing for a specific, conservative industry. Scored like 90% accuracy (above human for the segment, less if you would take the entire object into account). I encouraged them to keep contact with other students and the university, because I soon realized this would take atleast 4-10 3-month long thesis works that would need to build on eachother before getting really the market leading role they easily could achieve. I had basically no contact with the company, I just did it because I suggested it. Damn now I realize I marketed it to the wrong part of the company. Will need to reach out to the actual CTO of the mother-company. Damn, that is 3-4 lost years for them XD
@dragonsaige
@dragonsaige Жыл бұрын
It is generally true
@frazuppi4897
@frazuppi4897 Жыл бұрын
@@dragonsaige nope
@liutaurasbalnius
@liutaurasbalnius Жыл бұрын
Whipping and nae naeing simultaneously should be an Olympic sport
@PaulGreeve
@PaulGreeve Жыл бұрын
I thought you would have included the mobile processors from Apple, Samsung, Qualcomm, etcetera. They all include a Neural Processor these days but I never hear much discussed in relation to them, how powerful they are, what they’re actually able to do and so on. Given the many millions of people own phones that are powered by these processors, surely the potential to bring Edge AI to everyone is now here if they are used effectively.
@eafindme
@eafindme Жыл бұрын
Chips tradeoffs generality and specificity. General computing like CPU is Turing complete but it is bad for neural network, while neural processor or tensor core is specialized in accelerating neural network computing but not Turing complete. It is not a big deal usually, just that how far they could push the efficiency envelope matters for the IoT edge computing.
@pyromen321
@pyromen321 Жыл бұрын
@@eafindme this is not true. All modern tensor processing hardware on mobile devices (Qualcomm, Apple, Google) is literally just a super light arm core with heavy modifications. The neural processors in all modern phones are all just as Turing complete as the main CPUs, but they are directly in control of silicon that is highly specialized for doing common neural network operations efficiently. Plus, any hardware that can do common neural network operations is guaranteed to be Turing complete unless it’s some of that analog trash that won’t be viable for 50-400 years.
@eafindme
@eafindme Жыл бұрын
@@pyromen321 thanks for the rectification. Because I'm currently working on neuromorphic computing on FPGA, it is hard to say if it could not be viable. Who knows computing paradigm will change in the future.
@franks_9254
@franks_9254 Жыл бұрын
My concern is, that those chips are nor easily available, if you are designing an industrial edge device. Right? While the type of chips shown in the video are.
@StephenGillie
@StephenGillie Жыл бұрын
Game-changing MAC (multiply-accumulate) chips which can each replace whole racks of traditional CPUs. QMAC (quantum multiply accumulate) qubits are analog bits, giving sub-bit precision, and analog outputs cut training time by 10x or more. There are several innovations in this space going past arm64/x86-64, and many will be ideal for mobile devices, but none are mentioned in this video.
@refikbilgic9599
@refikbilgic9599 Жыл бұрын
I would like to write some personal sentences since this is something I am recently working on. There are already existing ASICs around, and it has a huge benefit compared to FPGAs, that is the "area". It's nice that we can reprogram our FPGAs anytime, but considering the edge-AI commercial devices (E.g., VR headsets), we have a very tight form factor for the chip. Then, it is feasible to say ASICs will dominate the edge-AI products. Even more, we will find the architecture of edge-AI, next to the processing units with more caches (everything on a single package = inside of our mobile devices in the future(?) ). However, we need to consider some issues with AI accelerators.. They have to work quite busy, which makes their thermal profile a bit annoying, and we need too many buffers next to cache memories (too much memory on-chip = too much area... ). We have nice cooling solutions already, but we definitely need more, or we need new methods to reduce the sparse computation of neural networks. Maybe you heard about "Spiking Neural Networks (SNN)". They provide a nice "event-based" structure to the network, which allows you to create "idle" states for your computation... That is already a nice idea to have a nice edge-AI chip with low power option !! Next, what if we make this chip in 3D? Considering the memory domination in the AI chips, what about stacking the memory die, vertically on top of the logic die? We try to answer this question in imec.
@severussin
@severussin Жыл бұрын
You work at Imec?
@refikbilgic9599
@refikbilgic9599 Жыл бұрын
@@severussin Yes… Specifically on this subject!
@diamondlion47
@diamondlion47 Жыл бұрын
Looks like AMD is already doing that with the mi300 line of chips. Wonder what your take would be on how much of a performance benefit this could give over an H100? Thanks.
@refikbilgic9599
@refikbilgic9599 Жыл бұрын
@@diamondlion47 AMD also work with imec, and indeed, not only AMD but other vendors are founding the R&D in this domain, such as Huawei, Google, Qualcomm and Intel. But these vendors mainly concentrate on the "development" process instead of "research". We can measure performance with total floating point operations per second (FLOPS), and both processors provide extremely high computation. But, as discussed, this is not enough, and we need to think about the power of consumption. In this case, one can measure the energy efficiency per operation cycle (FLOPS/Watt). But overall, performance measurement is not straight-forward and we can think about many other parameters like, logic core configurability (multi/many-core with multi-thread ops), memory subsystem (L1/L2 cache sizes, etc.), technology processing node etc... Finally, as the end user, we only see the marketing prices and advertised performance comparison for last products...
@diamondlion47
@diamondlion47 Жыл бұрын
@@refikbilgic9599 ​ @Refik Bilgiç Thank you very much for the response. I notice you don't mention Nvidia in the list of companies, do they partner with other research institutions or are they not doing as much research into stacking or advanced packaging etc or just an oversight?
@dominic.h.3363
@dominic.h.3363 Жыл бұрын
Natural language processing is what will end me. There is literally nothing else I can do as a paraplegic with dyscalculia on the ass-end of Europe besides translation, and Deepl is already near perfect between every language pair I can speak. I've got every recommendation under the sun from becoming a musician (been playing the guitar for 6 years daily, can't memorize the fretboard because of dyscalculia), to programming (can't memorize the syntax because of dyscalculia), to 3D modeling... (you get the gist, nothing even remotely related to manipulating quantities or memorizing arithmetic relations is viable), to becoming a novelist (sure, because we all know how great those people earn). Aanyway, that was my comment for the algorithm.
@brainletmong6302
@brainletmong6302 Жыл бұрын
Damn, god does curse twice indeed. Hopefully you'll figure something out soon. I'd suggest language teaching to students but I'm sure you've already thought about that one for a long time.
@AschKris
@AschKris Жыл бұрын
People who can't work shouldn't have to work, why are we advancing tech and automating jobs if we're not also reducing the necessary work to make society function?
@dominic.h.3363
@dominic.h.3363 Жыл бұрын
@@brainletmong6302 I can't teach because I don't know any grammar I don't "get" grammar. Guess why. I acquired every single language I am speaking by exposure to the language, much like a toddler learns to speak by just simply absorbing the language in practice from people who are speaking around them. I've learned my first foreign language with a starting vocabulary of ~300 words within 2 and a half months at eight years of age watching TV 18 hours a day, because I was bored with nothing else to do on my summer break. At the start of my summer I watched cartoons. At the end I watched nature documentaries. English followed as my second foreign language soon after... I'm not stupid, I have a very specific impediment that makes me unable to do a large category of things, and another impediment that makes me unable to do most of the things I have left after the first impediment. The tiny sliver that is left is being made superfluous as a field of employ by AI. At the same time I have lexical knowledge about several things I would be unable to perform. For instance, you could have me in any aquarist store and I would be perfectly capable of disseminating expert knowledge to prospective aquarists about both aquatic flora and fauna, but I could neither stock shelves, or get fish from tanks that are outside my reach.
@dominic.h.3363
@dominic.h.3363 Жыл бұрын
@@AschKris That's a commendable notion but I don't like to feel useless... which I do, increasingly so. I want to be able to succeed at something that gives me a sense of fulfilment. And it's not giving me the satisfaction if I am doing it for myself, I want to contribute to "making society function" as you put it.
@AschKris
@AschKris Жыл бұрын
@@dominic.h.3363 Yeah, but that's because you want to do it, it shouldn't be a requirement to stay out of the streets.
@rngQ
@rngQ Жыл бұрын
2:35 The AI-generated PCB image is a nice touch
@murdercom998
@murdercom998 Жыл бұрын
ty for your hard work =)
@jjhw2941
@jjhw2941 Жыл бұрын
Some of the Nvidia Jetson boards like the AGX Xavier and Orin have separate Nvidia Deep Learning Accelerators built in as well as the GPUS. There is the Hailo-8 M.2 accelerator too.
@CrazyEngineer314
@CrazyEngineer314 Жыл бұрын
You should take a look at Perceive's ERGO chip it seems to be a gamechanger in this field
@jamessnook8449
@jamessnook8449 Жыл бұрын
Almost 20 years ago I led a project to port the neural models we ran on a Beowulf cluster to a more mobile platform. Our goal wasn't to create a processor to run solved networks like the current crop of AI processors - we built it with full plasticity so that the learning and processing could be performed on the same piece of hardware. I am disappointed that what is available today is a shadow of what we did in 2004. None of the current processors are constructed to model the neural structures required for real intelligence. They just keep modeling the same vacuous networks I used as an undergrad in the cog sci program at UCSD in the late 80's. Most of the people using the technology don't understand intelligence and sadly don't care what is required. One example: Lately I've seen numerous job postings for AI engineers who can do prediction - what they don't understand is that it isn't prediction, the missing component of these networks is expectation - facilitated by prospective memory.
@abowden556
@abowden556 Жыл бұрын
There are some people who actually care about this, Numenta for example! They partnered with AMD/Xilynx, so it's not like their approach has no support. Sparsity i a huge win, but their plans go way, way beyond that. They actually want to model the brain, or a small fraction of it, and their research has already borne fruit in the area of sparsity. They are definitely wanting make hardware that more closely mirrors the brain, when they get the chance though. It's very clearly their primary area of interest.
@VenkataRahulSatuluri0
@VenkataRahulSatuluri0 Жыл бұрын
Are you speaking of probablistic learning sir?
@jamessnook8449
@jamessnook8449 Жыл бұрын
@@abowden556 Functionally appropriate neural processors aren't easy to design/build and won't be as efficient as current neural processors are at running solved networks, but the current methodology won't give rise to real intelligence either. Sparsity? I believe you're referring to what we called sparse connectivity - is one aspect of what is missing from the newest versions of what is really just improved back propagation networks. Another missing element is learning - even my professors thirty years ago admitted that there was no biological basis for back prop, but they had no other mechanism for modifying connection strengths. Few people are incorporating real biology in their network designs because it is a pain in the butt, and truthfully even fewer care about it. I am glad someone (Numenta you mentioned) still does.
@jamessnook8449
@jamessnook8449 Жыл бұрын
@@VenkataRahulSatuluri0 In regard to which portion of my comment are you referring?
@VenkataRahulSatuluri0
@VenkataRahulSatuluri0 Жыл бұрын
Sir, the last part, expectation facilitated by prospective memory
@Bareego
@Bareego Жыл бұрын
The main issue I have with AI is that we let silicon/software learn under some evolutionary pressure, and we pick the one that does the job best. But we don't actually understand what made that particular resulting structure do the job best. We harness the gains from complexity without learning about the part of the structure that makes this so efficient and why. It's like someone finding a wand and wielding it without knowing how it does what it does. Part of this is that more complex problems require an AI network that is so complex, that we have no hope of understanding what makes the best working model tick. I don't think we're in any danger of being taken over by AI, but that the information we could learn from AI for making our own designs are rarely or never learned. As edge AI is concerned, I suspect we'll get finished AI solutions preprogrammed. This leaves a small requirement of processing for the edge device, which should keep power requirements low. Most devices are VERY specialized in their purpose. Much easier to run an expensive AI centre and churn out cheap pre-programmed devices.
@droid16beta97
@droid16beta97 Жыл бұрын
Pre-programmed as opposed to what? They don't train the models on the go...
@vishalgiraddi5357
@vishalgiraddi5357 Жыл бұрын
And for that there's a topic called explainable AI which aims to predict not only th result, but also explain why the particular result was obtained, presently explainable AI models are not as powerful as conventional black box models, but they exist
@bakedbeings
@bakedbeings Жыл бұрын
Look into anaesthesia. We've been putting millions of surgery patients a year under deep chemical anaesthesia for decades without knowing the mechanism.
@tt10tt
@tt10tt Жыл бұрын
​@@bakedbeings excellent analogy.
@hypercube33
@hypercube33 Жыл бұрын
Love the potato PCB ai generated image at 2:44
@kennethtan6403
@kennethtan6403 Жыл бұрын
Thank you and much Love from the Philippines.
@MaxxPa1
@MaxxPa1 Жыл бұрын
Hello you have been selected among my lucky winners DM via the above name on telegram to claim your prize 🌲 🎁.
@Shaunmcdonogh-shaunsurfing
@Shaunmcdonogh-shaunsurfing 9 ай бұрын
Fantastic video
@mbabcock111
@mbabcock111 Жыл бұрын
Brainchip AkidaTM is the world’s first commercial neuromorphic processor. It mimics the brain to analyze only essential sensor inputs at the point of acquisition-rather than through transmission via the cloud.
@Hyo9000
@Hyo9000 Жыл бұрын
Fwiw, CUDA doesn't really use C or C++. It has a very restricted API from which you can call upon its features, and yes it is accessed from "C", but its semantics are very different. Branches are expensive, and recursion is not allowed. And malloc doesn't exist.
@alexandrep4913
@alexandrep4913 Жыл бұрын
Video summed up: There is no shortcuts, this is going to require technology we just dont have yet to really get what we want from AI/Deep Learning
@Stroporez
@Stroporez Жыл бұрын
What about in-memory compute and neuromorphic chips?
@LogioTek
@LogioTek Жыл бұрын
Qualcomm AI100 DM.2e form-factor edge AI accelerators scale 70-200 TOPS at 15-25W, also various Qualcomm Snapdragon mobile SoCs fit the role of less powerful 30 TOPS and under edge AI accelerators. Qualcomm is pushing pretty hard into that direction with SoCs for edge AI, vision AI, drones/robotics, ADAS, and level 4-5 autonomous driving including software stacks. They even have AI research division dedicated to solving problems with edge AI model optimizations and other things.
@ChrisSudlik
@ChrisSudlik Жыл бұрын
You're missing the latest development in edge AI: simplified models running on a physical quirk flash memory. Mythic AI and Syntiant are two companies taking advantage of this to do simple, this tech is in the earliest days and has a lot of future potential.
@bakedbeings
@bakedbeings Жыл бұрын
I've googled quirk flash memory to no avail.
@jagadishk4513
@jagadishk4513 Жыл бұрын
Veratasium made a video on mythic
@ChrisSudlik
@ChrisSudlik Жыл бұрын
@@bakedbeings a quirk in flash memory, I missed a word in there. Veratasium has a somewhat ok primer on the general principle.
@aatkarelse8218
@aatkarelse8218 Жыл бұрын
Must agree, training models is tough, using them much less so. a bit of a none issue vid.
@nailsonlandim
@nailsonlandim Жыл бұрын
I saw that one, and it promises!
@maus3454
@maus3454 Жыл бұрын
I was wondering whether it is possible to un-learn or de-learn (undo a learning) of a learned neural network? So it will forget things or patterns
@dragonsaige
@dragonsaige Жыл бұрын
This is definitely possible, for example by simply further training the model on something else, but why would you want that?
@maus3454
@maus3454 Жыл бұрын
@@dragonsaige In case it learned something you don't want it to
@monstercameron
@monstercameron Жыл бұрын
you didnt mention in-memory computing or analog computing as possible solutions. Take a look at Mythic AI, looks like the will get around the perf and cost by using an analog and in memory computing methodology to get around the von nueman bottleneck and use a flash memory manu process which is relatively cheap and dense.
@PhazerTech
@PhazerTech Жыл бұрын
Good point. Also the memristor is a breakthrough that could play an important role here too.
@slicer95
@slicer95 Жыл бұрын
Mythic AI has shut down.
@timng9104
@timng9104 Жыл бұрын
not sponsored but arduino nano 33 BLE with ARM M4F chip is so good, I have pretty good accuracies with some custom applications (shadow gesture sensing). There are also concepts on reservoir computing, recurrent networks, echo state networks, would love to hear your take
@jannegrey593
@jannegrey593 Жыл бұрын
This is gonna be interesting.
@NathanaelNewton
@NathanaelNewton Жыл бұрын
Heh I agree
@subliminalvibes
@subliminalvibes Жыл бұрын
This channel never fails.
@jannegrey593
@jannegrey593 Жыл бұрын
@@subliminalvibes agreed
@geoffegger
@geoffegger Жыл бұрын
Any thoughts on hailo?
@Maleko48
@Maleko48 Жыл бұрын
thanks for highlighting the NPU on a stick... might have to check that out
@glock7061
@glock7061 Жыл бұрын
The thing with fpga is that you have to manually program entire network in vhdl or something similar. Which is cumbersome and not very easy. Belive me I tried.
@ristekostadinov2820
@ristekostadinov2820 Жыл бұрын
don't worry, he mentioned that in the video 😄
@gorak9000
@gorak9000 Жыл бұрын
Sounds like you need a script to take your model and generate the VHDL automatically. I've done it in the past - VHDL doesn't have the automation you need sometimes to do what you need, so write an external script (c/c++/python, whatever) that generates the VHDL you need that's too tedious to do by hand.
@eafindme
@eafindme Жыл бұрын
I made a CNN to FPGA microarchitecture hardware transcription automtion software, so I only have to focus on CNN model and design. Of course, it is on exotic computing domain so not a typical binary computing.
@glock7061
@glock7061 Жыл бұрын
@@eafindme What about something more complex than CNN? Self attention for example
@illeatmyhat
@illeatmyhat Жыл бұрын
It seems more sensible to program the network in something like Amaranth (previously nMigen) in Python.
@eelcohoogendoorn8044
@eelcohoogendoorn8044 Жыл бұрын
'FPGAs have a lot of potential'; if I got a dollar for every decade thats been said... id be able to buy a fancy cup of coffee, at least. Not sure if FPGAs have some kind of massive marketing campaign, or if the general idea just appeals to nerds; its like FPGAs are to EEs as what fusion is to physicists. To present FPGAs as having even a fraction of a precent of the relevance to edge AI as jetson nano or GPU type architectures does, makes presenting ITER as a practical solution to power generation look intellectually honest and well informed by comparison.
@isaidromerogavino8902
@isaidromerogavino8902 Жыл бұрын
that's pretty harsh, but not gonna fight with you on the fpga issue, never used them. On the other hand, I do have my reservations about the feasibility of the ITER project, as well. However, I'll admit the sentiment is, in part, simple intuition derived from a bare understanding of physical phenomena/concepts. The other part of the sentiment coming from a (comparatively) more rigorous mathematical formation at school. Would you be so kind to guide me towards some source or reference to complement my knowledge on the theoretical/technical difficulties which inform your opinions on ITER?
@eelcohoogendoorn8044
@eelcohoogendoorn8044 Жыл бұрын
@@isaidromerogavino8902 This is a fairly recent decent video on the issues; though I think she is still being too kind: kzfaq.info/get/bejne/grBkipSYkpyumY0.html&ab_channel=SabineHossenfelder
@napalmholocaust9093
@napalmholocaust9093 Жыл бұрын
I have "The Edge" open in a back tab. Yt free movies has it up now. Or did, they only stay free for a few weeks sometimes and when I get to it its gone and private. I had to sit through "Rob Roy" dozens of times myself in earlier days.
@zyzzfanzyzzfan
@zyzzfanzyzzfan Жыл бұрын
I’ve been studying neural architecture search for my PhD but ive not heard much about joint architecture and hardware search. Currently in an internship in industry and that sounds like quite an ambitious goal based on whats going on meta.
@bobdobs23
@bobdobs23 Жыл бұрын
Am I mistaken is not Tesla with their dojo configuration among the leaders?
@J_X999
@J_X999 Жыл бұрын
Video on US chip restrictions?
@odopaisen2998
@odopaisen2998 5 ай бұрын
> Silicon Compiler : kzfaq.info/get/bejne/fbNpg65yy7rQnYU.htmlsi=b7se3OuI42jfNYkM > Parallella : kzfaq.info/get/bejne/rLxpmcaiurvJYY0.htmlsi=2dtIp--sL6L4iiKP&t=830
@alecdvor
@alecdvor Жыл бұрын
Is there an rss feed for the news letter
@MaxxPa1
@MaxxPa1 Жыл бұрын
Hello you have been selected among my lucky winners DM via the above name on telegram to claim your prize 🌲 🎁.
@larryslobster7881
@larryslobster7881 Жыл бұрын
more ai vids love this
@Embassy_of_Jupiter
@Embassy_of_Jupiter Жыл бұрын
Pretty sure the future is completely distributed compute, be that edge or cloud, everything will be automatically optimized to run as low cost/latency via markets. You'll be able to buy and sell the tiniest useful sliver of computation and your devices will automatically buy/sell from/to the lowest/highest bidder. So in the end there will be real time market forces shaping which computing devices will be where. If you have some large computation that isn't latency sensitive, it'll automatically be outsourced to the cloud, taking into account the size of the problem, the time, the latency, the size of the output. Even how to most cheaply get the end result back to you will be market driven, e.g. what's cheaper, storing it in the cloud, only accessing singular elements or should it be compressed and sent to you or is compression too expensive vs sending it uncompressed etc.. In areas of high computation costs/latency that will locally encourage the aggregation of computational power. You might get recommendations how to save money by buying/renting edge devices or to make money by increasing your local area's computational power, similar to miners. All based on the buying and selling of tiny packages of computations. And I suspect edge will definitely have it's place in that future. Maybe in the far future there's even drones that fly computational power from place to place wherever the highest prices are at the moment, like pidgeons looking for food. Some scientist starts a calculation and a giant swarm of drones flocks to his house, crawling over each other like a bee hive. Who knows lol.
@UnderscoreZeroLP
@UnderscoreZeroLP Жыл бұрын
In the future my poop will be able to recycle into food i think
@charlesfowler4308
@charlesfowler4308 Жыл бұрын
@@UnderscoreZeroLP That's nothing, mine will develop limbs and will farm and grow my food for me
@franklinhaut
@franklinhaut Жыл бұрын
Valeu!
@cryptoinside8814
@cryptoinside8814 Жыл бұрын
So AI computing is nothing more than a super-fast computer either sitting at the data center or edge. It's like a super IBM PC vs IBM mainframe in the old days 🙂 People just love to put lipstick on a pig and call it something different, but it's actually the same sh*t 🙂
@severussin
@severussin Жыл бұрын
Basically
@temistoclesvaldes5312
@temistoclesvaldes5312 Жыл бұрын
O man i got this 18 minutes fresh? nice
@jairo8746
@jairo8746 Жыл бұрын
5:35 It is really silly to think you can try pruning something that nobody has an idea of how or even why it works, to make it work more efficiently. I imagine it would be similar to cutting cables in an electrical panel and checking if the machine powered by it is working faster or slower, without knowing which room or component had had the power cut out.
@alefratat4018
@alefratat4018 Жыл бұрын
Yes, pruning is one of the most overrated approach when it comes to optimizing the accuracy-performance trade-off of neural networks. But because this idea came from renowned gurus of the DL field, lots of people think it's the way to go. In practice, it very often leads to disappointing results.
@dragonsaige
@dragonsaige Жыл бұрын
On many popular models pruning can give gpu class performance on cpus with only a few percent accuracy drop on modern inference engines.
@droid16beta97
@droid16beta97 Жыл бұрын
I don't feel like any of those are AI-specific issues but apply to all kinds of computation.
@FobosLee
@FobosLee Жыл бұрын
wait. Isn’t it an AI generated image at 2:40 ?
@sunnohh
@sunnohh Жыл бұрын
I work with NLU and it is almost all snake oil
@ChrisHarmon1
@ChrisHarmon1 Жыл бұрын
My father LOVES the same movie. Must be a dad thing.
@jadeaffenjaeger6361
@jadeaffenjaeger6361 Жыл бұрын
The edge NPU with arguably the best chances of widespread adoption are ARM's Ethos-U and Ethos-N NPUs which are made to complement their Cortex-M and Cortex-A processors respectively. To my knowledge they do not exist in (commercial) silicon yet but will like have a head start over the offerings of small startups due to their seamless integration with an already established ecosystem.
@NathanaelNewton
@NathanaelNewton Жыл бұрын
4:50 *realizes that I have been using premier pro the hard way*
@cjmitz
@cjmitz Жыл бұрын
I’m looking at using multiple edge ai devices ($3) in parallel. Each running a real time algo and then combining outputs. NNs will get progressively more efficient and will easily run on modern micros. Micros are the future over fpgas but an fpga can form part of the parallel to serialising etc.
@yelectric1893
@yelectric1893 Жыл бұрын
What’s a micro?
@smooooth_
@smooooth_ Жыл бұрын
@@yelectric1893 I'm assuming short for microcontrollers
@yelectric1893
@yelectric1893 Жыл бұрын
@@smooooth_ ahh. Wow, well having a dozen picos and communicating the results could be pretty robust
@tobuslieven
@tobuslieven Жыл бұрын
Maybe ASICs for standard image processing layers like edge and depth detection, where a lot of the hard work can be done efficiently and in a standard way, then FPGAs or GPUs for more specific neural nets that take the final layer from the standard ASICs and do something interesting with them.
@VenkataRahulSatuluri0
@VenkataRahulSatuluri0 Жыл бұрын
And then there came software programmable ASICs to run inferencing for generic AI models
@AscensionNow1
@AscensionNow1 Жыл бұрын
thanks
@hushedupmakiki
@hushedupmakiki Жыл бұрын
The way you close your video sounded exactly like my highschool coaching tutor. He would give really in depth and well explained lessons, but the hour ended, closed shop and got out in his car before the class' bags were even packed 🤣
@FerranCasarramona
@FerranCasarramona Жыл бұрын
Another edge AI techonolgy is spiking neural networks, which can reduce consumption a lot. Brainchip is a company offering this technology
@cavemaneca
@cavemaneca Жыл бұрын
Using edge processors for neural networks at work, it's amazing to see how quickly the field is growing and how quickly their capabilities grow.
@sodasoup8370
@sodasoup8370 Жыл бұрын
Honestly i cant see how this problem could be "solved". on an Edge device like for traffic recognition you need fast AND precise Inference. Like the Video said, most ways of cutting corners just dont apply. I fear that this will result in overspecialized hardware which will flood the market after some cheap product with good software support gets released...
@alefratat4018
@alefratat4018 Жыл бұрын
There is a common misconception (encouraged by the big tech companies for good reasons) that says that using bigger and bigger neural networks is always better. That's true only if you want to solver very generic problems (such as the ones tackled by those big tech companies). However, a lot of real-live scenarios and use-cases are much narrower and could be addressed by a set of specialized lightweight, cleverly designed, neural networks.
@sodasoup8370
@sodasoup8370 Жыл бұрын
@@alefratat4018 i am just starting to learn ML, but isnt it in the case of Computer Vision just a bunch of Convoluted layers which take 90% of the computing Power? The fully connected layers after that are important for the model but you cant save much resources with them...
@vishalgiraddi5357
@vishalgiraddi5357 Жыл бұрын
Interesting topic indeed, My own final year college project is based on post training quantization of models Currently working on performing 2 bit quantization
@bluegecko6770
@bluegecko6770 Жыл бұрын
Thanks for your enjoyable content
@goodfortune6399
@goodfortune6399 Жыл бұрын
That was a good movie
@jonathanpeace3167
@jonathanpeace3167 Жыл бұрын
Its all about uJ/inference and low Iq; checkout Syntiant and Greenwaves for example...
@Erik-gg2vb
@Erik-gg2vb Жыл бұрын
Will be interesting to see how Musk's, not sure if it is apples, DOJO works out.
@matneu27
@matneu27 Жыл бұрын
Learned a lot more thanks to your video and (I assume) the ai generated voice. Anyway ai could maybe one day optimize its power consumption and manage a power grid to optimize the distribution.
@EyesOfByes
@EyesOfByes Жыл бұрын
0:49 It's the dude from Ireland. Duh ;)
@MaxxPa1
@MaxxPa1 Жыл бұрын
Hello you have been selected among my lucky winners DM via the above name on telegram to claim your prize 🌲 🎁.
@samgeorge4798
@samgeorge4798 Жыл бұрын
Love your use of AI generated images throughout the video.
@artscience9981
@artscience9981 Жыл бұрын
As an electrical engineer, a video like this is music to my ears.
@MaxxPa1
@MaxxPa1 Жыл бұрын
Hello you have been selected among my lucky winners DM via the above name on telegram to claim your prize 🌲 🎁.
@computerconcepts3352
@computerconcepts3352 Жыл бұрын
interesting 🤔
@amandahugankiss4110
@amandahugankiss4110 Жыл бұрын
Okay I love you goodbye!
@johnshaff
@johnshaff Жыл бұрын
Power not processing is the problem
@arpinfidel
@arpinfidel Жыл бұрын
a topic i'm really interested in is the use of analog computers for NNs. most NNs are not entirely precise by nature so I think it would make sense to run it with analog chips
@MaxxPa1
@MaxxPa1 Жыл бұрын
Hello you have been selected among my lucky winners DM via the above name on telegram to claim your prize 🌲 🎁.
@RonJohn63
@RonJohn63 Жыл бұрын
0:06 Eat mor chikin
@pyromen321
@pyromen321 Жыл бұрын
I’m amazed you didn’t mention Qualcomm, Apple or Google. All of them have very good edge AI solutions compared to everything you mentioned in this video. Whenever you open up your phone’s camera app, data is being fed to multiple neural networks for autofocus, lighting, super resolution, noise reduction… etc. Qualcomm even has their AI hardware exposed to app developers. Not sure about apple and Google, though.
@nomadhgnis9425
@nomadhgnis9425 Жыл бұрын
Optical cpu🤔
@arubaga
@arubaga 6 ай бұрын
We should demand open source firmware, or edge AI could be insecure at brain stem level. 😮
@first-thoughtgiver-of-will2456
@first-thoughtgiver-of-will2456 Жыл бұрын
Constrained (edge) models will perform well once we have pretrained ANN based online reinforcement learning algorithms that solve discretization (episode length and SAR map resolution) and search strategy (not epsilon decay, a better bellman equation). I also don't think parameter quantization is reliably stable enough for chaotic systems (deep nets) and stochastic dropout mine as well be Monte Carlo architecture search. Quantization is a useful techniques but I don't think its the future of edge learning and is a nascent attempts at porting big data to the edge instead of finding an edge first solution or integration to deep net models. Inference at the edge is underutilizing the edge. We need fine tuning at the edge and continual learning. Tesla's closed loop data feedback and update pipeline is an interesting approach that utilizes the constantly shifting domain seen by the model at the edge.
@Smitty_Werbenjagermanjensen1
@Smitty_Werbenjagermanjensen1 Жыл бұрын
I don’t think I’ve ever heard of the terminology ‘edge’ devices, I believe they’re called IOT (internet of things) devices in Cisco and maybe elsewhere
@ryanreedgibson
@ryanreedgibson Жыл бұрын
You need to start charging for this. By no means stop posting.
@deusexaethera
@deusexaethera Жыл бұрын
Today I learned Github Copilot exists.
@MaxxPa1
@MaxxPa1 Жыл бұрын
Hello you have been selected among my lucky winners DM via the above name on telegram to claim your prize 🌲 🎁.
@RonJohn63
@RonJohn63 Жыл бұрын
The real problem is that computer neural networks don't work like biological neural networks.
@jasonwaterfalls2935
@jasonwaterfalls2935 Жыл бұрын
Why not offload computing to an edge cloud? This combined with 5G would solve the latency constraint.
@alefratat4018
@alefratat4018 Жыл бұрын
Not possible for scenarios when you have to cope with hard real-time constraints. Which are quite common actually.
@again5162
@again5162 Жыл бұрын
I hope we eventually destroy disease with better computers especially cancer
@MaxxPa1
@MaxxPa1 Жыл бұрын
Hello you have been selected among my lucky winners DM via the above name on telegram to claim your prize 🌲 🎁.
@daviddevlogger
@daviddevlogger Жыл бұрын
Second
@JeffreyCC
@JeffreyCC Жыл бұрын
Why do I always get you in my recommendations, also when I blocked you in every way possible?
@alex15095
@alex15095 Жыл бұрын
Why did you block him?
@JeffreyCC
@JeffreyCC Жыл бұрын
@@alex15095 Because his video's were occupying a third of my daily feed without watching any of them
@alex15095
@alex15095 Жыл бұрын
@@JeffreyCC KZfaq algorithm 👍
@styx1272
@styx1272 Жыл бұрын
What about the Akida 1000 by Brainchip Corp ? You seem to have missed a winning entry. And deleted my post ?
@TheMoeShun
@TheMoeShun Жыл бұрын
tinygrad
@GavinM161
@GavinM161 2 ай бұрын
Good lord, what tosser bought a license plate with "TSLA S1"!?!
@irvingchies1626
@irvingchies1626 Жыл бұрын
Deep learning ≠ A.I. Actual A.I. should be able to learn just by watching once or twice, then a little practice just like a human
@jmatx
@jmatx Жыл бұрын
Edge, edge, the edge, the Edge -- what is the Edge, edge?
@emmybrown4031
@emmybrown4031 Жыл бұрын
Amazing video and thank
@JohnWaynep1
@JohnWaynep1 Жыл бұрын
That's awesome. I dealt with crypto last year on Robinhood, tried some index but didn't take it out so I lost it by the end. Any consistent strategies?
@elizabethholman9154
@elizabethholman9154 Жыл бұрын
She is also my personal trader, crypto analyst and account manager. With an initial invested capital of $8000, it yielded returns of over $22000 within two weeks of trading. I was really impressed by the profit Actualized
@liciaramos320
@liciaramos320 Жыл бұрын
@@elizabethholman9154 Same here, I started with $3,000 now earning $28,000 bi-weekly profits with her trading program.
@liciaramos320
@liciaramos320 Жыл бұрын
@Sarah Helen l'm so sure! She's very active on What's Apk"
@liciaramos320
@liciaramos320 Жыл бұрын
She’s active with this +𝟭𝟱𝟰𝟭𝟲𝟯𝟴𝟴𝟵𝟱𝟮 currently
@reddixiecrat
@reddixiecrat Жыл бұрын
Sounds like you’ve never heard of Palantir
@brettyoung6045
@brettyoung6045 Жыл бұрын
The reason there aren’t that many good edge chips is because the market is small.
@coraltown1
@coraltown1 Жыл бұрын
I am really not all that impatient for ultra powerful AI to appear on The Edge, as so much of this tech gets misused. Generally, it's just not making the world a better place.
@V3racious3
@V3racious3 Жыл бұрын
If I hear about the Asianometry Newsletter one more time today I'm actively spending resources against it.
@nailsonlandim
@nailsonlandim Жыл бұрын
Great video @asiaonometry! Iḿ suspect but I tend to favor NVidia basically because is easier to deliver the models and inference code, basically what changes is the Docker container and a few libraries. Besides that flexibility comes at a cost of power consumption. For instance if I want them running 24x7 on solar independent power I need a large solar panel AND lot's of batteries
@tsunamio7750
@tsunamio7750 Жыл бұрын
thanks
AI’s Hardware Problem
16:47
Asianometry
Рет қаралды 618 М.
Why Moore’s Law Matters
23:41
Asianometry
Рет қаралды 237 М.
Haha😂 Power💪 #trending #funny #viral #shorts
00:18
Reaction Station TV
Рет қаралды 14 МЛН
ИРИНА КАЙРАТОВНА - АЙДАХАР (БЕКА) [MV]
02:51
ГОСТ ENTERTAINMENT
Рет қаралды 4,2 МЛН
Small Brain, Big Think: AI on the Edge
12:40
Zack Freedman
Рет қаралды 454 М.
How AI 'Understands' Images (CLIP) - Computerphile
18:05
Computerphile
Рет қаралды 178 М.
The Magic of RISC-V Vector Processing
16:56
LaurieWired
Рет қаралды 215 М.
“The Decision of the Century”: Choosing EUV Lithography
24:28
Asianometry
Рет қаралды 226 М.
Popular Technologies that Won't be Around Much Longer...
14:36
Sideprojects
Рет қаралды 113 М.
Why AMD's Chiplets Work
12:53
Asianometry
Рет қаралды 292 М.
AI Hardware w/ Jim Keller
33:29
Tenstorrent
Рет қаралды 28 М.
This is NOT a Graphics Card - ASUS AI Accelerator
15:24
Linus Tech Tips
Рет қаралды 1,9 МЛН
The Chips That See: Rise of the Image Sensor
18:29
Asianometry
Рет қаралды 156 М.
Where The Real Chip Shortage Is
11:53
Asianometry
Рет қаралды 614 М.
Haha😂 Power💪 #trending #funny #viral #shorts
00:18
Reaction Station TV
Рет қаралды 14 МЛН