--
You received this message because you are subscribed to the Google Groups "HomeBrew Robotics Club" group.
To unsubscribe from this group and stop receiving emails from it, send an email to hbrobotics+...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/hbrobotics/8a36d910-7a48-4f2b-8d14-c12b39f0bec8n%40googlegroups.com.
The 4-bit modes are used to hold gigantic models in much less
memory. The lower the number of bits per 'float', the less
fidelity, but sometimes it seems to work close to the same. Maybe
some of these models have variable numbers of bits for different
parts of the model? (Want to dig into this soon...)
The GPUs use very little power when not active. My liquid cooled RTX-4090 very clearly powers up the cooler when stressed to various levels, then is silent otherwise - a pretty good measure of power usage.
Pretty sure we're going to learn how to compress networks down to be much more compact soon. Of course we'll use that to create effectively even larger networks. But we should get some very useful aspects to fit nice size / power envelopes. When we settle on specific network models, and even specific weights to some extent, we can head toward the ASIC and/or custom processor architectures to get power down to almost nothing. This is what the Movidius chips did for camera image processing: gigantic bandwidth, good per-pixel processing, on milliwatts.
In the mean time, a heavy-duty PC is going to be expensive & power hungry...
Seems like the main players are: NVidia, Apple, Qualcomm, Intel, and AMD. Qualcomm is probably going to keep the lead for low-power high function mobile, with NVidia & Apple fighting out the very high end + low power, i.e. Low-SWaP. Apple is pretty good on mobile, but not on broadly usable mobile: NVidia & Qualcomm have that end locked down.
Streaming pose + camera views + sensing to a cloud brain to get
control inputs is viable for a certain range of things. Having a
local control loop that can sample + control in less than the
20-200ms range for wireless network roundtips that can still be
guided by the cloud brain might be the best combination.
sdw
The new iPad Pro with the M4 chip can provide 38 trillion operations per second at low power, half the wattage of the previous version.
It supports Thunderbolt 4 (and USB4) which provides 32Gbps of PCIe. Theoretically, you could connect Jetsons, Snapdragons, and even some microcontrollers to one or more iPad Pros (etc.) over PCIe at up to 32Gbps at low power. And have some nice displays, removable 'brain'.
What do you know (err, wa-d'ya-know), there is already an open
source board to do Thunderbolt<->PCIe: (And, separately,
10GigE interfacing on the Jetsons via M2.)
https://github.com/antmicro/thunderbolt-pcie-adapter
sdw
|
Stephen D.
Williams
Founder: VolksDroid, Blue Scholar Foundation |
--
You received this message because you are subscribed to the Google Groups "HomeBrew Robotics Club" group.
To unsubscribe from this group and stop receiving emails from it, send an email to hbrobotics+...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/hbrobotics/e198e08f-82a2-435f-a75f-a29802054331%40lig.net.
A Mac mini is nice, less expensive because it doesn't have a display, battery, or mobile efficiency. But on-robot compute could benefit from those, if it is worth including an expensive brain.
The new iPad Pro M4 tablets have up to 16GB LPDDR5X RAM.
https://www.trendforce.com/news/2024/05/08/news-apple-unveiled-m4-chip-for-ai-heralding-a-new-era-of-ai-pc/
I'm not clear how much of a pain it would be to write an app to make use of the GPU, neural chip, etc. A Mini is going to be a lot easier to develop for. Probably best to develop for the Mini, then deploy somewhat widely on the iPad etc.
I'm a lot more comfortable developing for Android, so any Android device is good there. Especially if you can root it, but that isn't so required anymore.
The Samsung S24 Ultra has 12GB of RAM. In a few years, refurb
S24s are going to make great robot computers. I have one old
Samsung for just that reason.
sdw
To view this discussion on the web visit https://groups.google.com/d/msgid/hbrobotics/A6EE4695-2E41-4609-8965-1A761F9FD11D%40gmail.com.
On May 10, 2024, at 6:43 PM, 'Stephen Williams' via HomeBrew Robotics Club <hbrob...@googlegroups.com> wrote:A Mac mini is nice, less expensive because it doesn't have a display, battery, or mobile efficiency. But on-robot compute could benefit from those, if it is worth including an expensive brain.
The new iPad Pro M4 tablets have up to 16GB LPDDR5X RAM.
https://www.trendforce.com/news/2024/05/08/news-apple-unveiled-m4-chip-for-ai-heralding-a-new-era-of-ai-pc/I'm not clear how much of a pain it would be to write an app to make use of the GPU, neural chip, etc. A Mini is going to be a lot easier to develop for. Probably best to develop for the Mini, then deploy somewhat widely on the iPad etc.
I'm a lot more comfortable developing for Android, so any Android device is good there.
To view this discussion on the web visit https://groups.google.com/d/msgid/hbrobotics/b3fe2508-fed1-4e66-ab8f-e9ee782e755f%40lig.net.
On May 11, 2024, at 8:17 PM, A J <aj48...@gmail.com> wrote:The recent CPU & GPU chips are very fast but software and AI might run faster if thechips were 10 - 15x bigger.
To view this discussion on the web visit https://groups.google.com/d/msgid/hbrobotics/55ff2e29-bbcd-48b9-be34-8be57355fbefn%40googlegroups.com.
I looking at the openXLM source code and it appears to run on Linux/Nvidia or an iPhone or Mac. From comments in the code it seems they are still training on Linux-on-Intel
--
You received this message because you are subscribed to the Google Groups "HomeBrew Robotics Club" group.
To unsubscribe from this group and stop receiving emails from it, send an email to hbrobotics+...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/hbrobotics/9579c7e6-c0e3-4e3a-95b3-5f62f134d9ff%40lig.net.
On May 20, 2024, at 10:05 AM, A J <aj48...@gmail.com> wrote:The compute power for robots seems to have a pretty good path for the next decade.If industry could support quality open source collaborations across hardware platformsit would help. With so many programmers in the world what Dev Model would work best?
On Sunday, May 12, 2024 at 1:23:42 PM UTC-7 Chris Albertson wrote:Sorry a typo, openELM. (ELM = Efficient Language Model).This is Apple’s latest, their goal was not to be smarter than GPT-4 but to be MUCH faster and run locally on an iPhone or Mac. It can also run on an Intal PC nut you’d want to have a mid-range Nvidia GPU card.Apple claims that openELM is comparable to Meta’s Llama2 but it uses less than half as many parameters and took much less time to train. If this is true, I’ll likely end up runing openELM. But right now my goal is to set up a searver to runs openAI’s API. The server will be able to use different LLMs. To make my job easier I’m initially going to use Llama2 because everything aroubnd it is more mature and seems better documented. I can switch later.Runing an openAI API server locally will be a resource for any future robot and I like the idea of a stable API. It’s not going to be finished today, there is much reading and I have to track down many dependencies and make each of them work.On May 12, 2024, at 12:09 PM, 'Stephen Williams' via HomeBrew Robotics Club <hbrob...@googlegroups.com> wrote:On 5/10/24 10:07 PM, Chris Albertson wrote:I looking at the openXLM source code and it appears to run on Linux/Nvidia or an iPhone or Mac. From comments in the code it seems they are still training on Linux-on-IntelopenXLM?--
You received this message because you are subscribed to the Google Groups "HomeBrew Robotics Club" group.
To unsubscribe from this group and stop receiving emails from it, send an email to hbrobotics+...@googlegroups.com.To view this discussion on the web visit https://groups.google.com/d/msgid/hbrobotics/9579c7e6-c0e3-4e3a-95b3-5f62f134d9ff%40lig.net.--
You received this message because you are subscribed to the Google Groups "HomeBrew Robotics Club" group.
To unsubscribe from this group and stop receiving emails from it, send an email to hbrobotics+...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/hbrobotics/9bd8c5fe-3855-4fa5-a1eb-1607e1be9288n%40googlegroups.com.
On Mon, May 20, 2024 at 4:49 PM, 'Dan' via HomeBrew Robotics Club<hbrob...@googlegroups.com> wrote:
On May 20, 2024, at 2:11 PM, camp . <ca...@camppeavy.com> wrote:> But by all means...keep drinking the Linux/ROS2 Kool Aid.If you're waiting on me to write a mapping, localization, navigation, or visualization routine forget about it. But with ROS I am able to copy and paste my way. It's not so much laziness as just getting it done plus interacting with folks who know a lot more than me; speaking a common language.
To view this discussion on the web visit https://groups.google.com/d/msgid/hbrobotics/2060757349.2226210.1716239469260%40mail.yahoo.com.
But on another topic....4 bit ai...I have heard a lot of buzz lately about custom hardware (ASICs) that are optimized for specific LLMs.
I am also uncomfortable piling on thick, complex, inefficient
layers just because it is easy to do so. It's the general problem
I have with a lot of Java stacks & approaches. ROS2 still has
me a bit uncomfortable, although it has a great argument for use
in prototyping, trying out drivers to new hardware. I suspect
that for anything that gets past the prototype stage, I'll want to
build something lean that is equivalent. I'm unclear how
practical that is in varous scenarios, but I know I'll want to.
On the OS, I guess it depends on a lot of factors. Linux is
always going to have the best ease of development, deployment,
etc. Zephyr is probably a good target for embedded compute above
extremity controllers.
NVIDIA Jetson boards can be configured to boot in 3 seconds:
https://docs.nvidia.com/jetson/archives/r34.1/DeveloperGuide/text/SD/Kernel/BootTimeOptimzation.html
Depending on details, it should be possible to boot Linux in
under a second. Some Android phones could boot very quickly,
although radio startup often takes a bit after that.
There is a major Linux-like, Linux-inheriting embedded OS for
more hard-core real-time & fast startup cases: Zephyr. It is
now a Linux Foundation project, and is being supported by more and
more hardware manufacturers. My impression of it is that it
borrows as much as possible from Linux to solve hard problems, but
is directly focused on being a good RTOS. Probably the most
important thing borrowed is the device tree & device drivers.
I suspect it overlaps the Linux ABI quite a bit.
https://www.zephyrproject.org/
"Zephyr supports more than 600 boards. Search our list for the hardware used in your application. This diversity of supported boards gives developers and product manufacturers multiple options to solve their embedded RTOS challenges with Zephyr. If your board is not supported out of the box, adding support for a new board is simple."
Zephyr has micro-ROS, a port of ROS2:
I happened to find Zephyr when I found civetweb and was integrating it along with many other C/C++ libraries into a CMake build. I saw a reference to Zephyr in the civetweb CMakeLists.txt. Zephyr is a whole OS as a single CMake build. Normally a cross-compiling build I expect. That's very nice.
I'm comfortable stripping down Linux, or using already stripped
down Linux instances. I used to regularly reconfigure, rebuild,
and modify the Linux kernel. A number of alternatives to the
monolithic kernel have been created, but through a number of
clever techniques, the Linux kernel has mostly keep up and been
better. But it does take a certain level of complexity, at least
to run fast. There are some fun projects that run Linux kernels
in an emulator on hardware that doesn't have an MMU, using
external storage to get enough memory. Not practical, but neat.
However, with Zephyr, I don't expect to be doing such Linux
hacking. And powerful enough microcontrollers are inexpensive
enough that things like uclinux are probably not worth the effort.
https://www.eetimes.com/how-uclinux-provides-mmu-less-processors-with-an-alternative/
https://popovicu.com/posts/789-kb-linux-without-mmu-riscv/
This is an interesting article on related topics: https://jaycarlson.net/embedded-linux/
sdw
To view this discussion on the web visit https://groups.google.com/d/msgid/hbrobotics/222523708.2211334.1716238184776%40mail.yahoo.com.
Using Linux as an example might not be the best way to start an open source project...Read the Wiki history of Linux. Originally Linus created a kernal, not a full OS. He ported other peoples GNU code to get the kernal to do something. He also kept the original kernal under his own license for a few years.Additionally, IMHO, an robot OS should not take 30 seconds to boot up. Imagine a robot pilot that for any number of reasons reboots while landing.
OK for a prototype perhaps, but in general, horrifying. In just
about all circumstances, communications should take at most 3
seconds, and usually <1s. OS startup should be fast, although
to be fully up might take longer to do full system checks, scans,
validations. There are a lot of bad messaging systems out there,
often connected to convoluted cloud computing systems.
My 2023 Toyota Highlander Hybrid is ready to shift into reverse
in about 1-2 seconds from when I hit the start button. I know it
is ready when it unlocks shifting. I do all of my fiddling after
I've started rolling. The entertainment system takes a few more
seconds to start up. It isn't really booting of course, just in
power save mode with the system computer always mostly ready. A
lot like a mobile phone. Robots, generally, will be always-on
like a mobile phone, table, or computer. MacBooks are always
running, even when mostly sleeping: They wake up periodically in a
very low power, low clock rate mode to check email, etc. (I had a
friend who managed the email group at Apple one time who mentioned
that feature.)
sdw
--
You received this message because you are subscribed to the Google Groups "HomeBrew Robotics Club" group.
To unsubscribe from this group and stop receiving emails from it, send an email to hbrobotics+...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/hbrobotics/6f4c5c21-f468-4b63-9242-6fd8dac8e4a1n%40googlegroups.com.
On May 21, 2024, at 2:04 PM, 'Dan' via HomeBrew Robotics Club <hbrob...@googlegroups.com> wrote:I think there is a micro-ROS port to FreeRTOS.