Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

- Motion planning: already discussed.

- Multiaxis singularities: much less of a problem than it used to be. We don't need closed-form solutions any more; we have enough CPU power at the robot to deal with this. You need some additional constraint, like "minimize jerk" when you have too many degrees of freedom.

- Simultaneous Location and Mapping. SLAM for short: Getting much better. Things which explore and return a map are fairly common now. LIDAR helps. So does having a heading gyro with a low drift rate. Available commercially on vacuum cleaners.

- Lost Robot Problem: hard, but in practical situations, markers of some kind, visual or RF, help.

- Object manipulation and haptic feedback: Look up the DARPA manipulation challenge. Getting a key into a lock is still a hard problem. It's embarrassing how bad this is. Part of the problem is that force-sensing wrists are still far too expensive for no good reason. I once built one out of a 6DOF mouse, which is just a spring-loaded thing with optical sensors. Something I was fooling around with before TechShop went down. I have a little robot arm with an end wrench and a force sensing wrist. The idea was to get it to put the wrench around the bolt head by feel. Nice problem because a 6DOF force sensor gives you all the info you can get while holding an end wrench.

- Depth estimation: LIDAR helps. The second Kinect was a boon to robotics. Low-cost 3D LIDAR units are still rare. Amusingly, depth from movement progressed because of the desire to make 3D movies from 2D movies. (Are there still 3D movies being made?)

- Position estimation of moving objects: the military spends a lot of time on this, but doesn't publish much. Behind the staring sensor of a modern all-aspect air to air missile is - something that solves that problem.

- Affordance discovery: Very little progress.

The real problem: solve any of these problems, make very little money. If you're qualified to work on any of these problems, you can go to Google, Apple, etc. and make a lot of money solving easier problems.



"The real problem: solve any of these problems, make very little money" - Just curious why have you come to this conclusion ?

Object manipulation has potential products in dishwashing and vegetable chopping - sufficiently large markets, potential billion $ outcomes for a startup which takes the early mover lead. Two robotic hands that can work in co-ordination just as human hands do. Extremely difficult to solve, but money is there.


I'm not the OP, but the real issue in robotics is two-fold and he mention both.

1. The cost & technical limitation. The cost, as some of the solution per problem OP suggest include "Lidar, visions and sensor" and all of those thing cost a lot by themselves. On top of that you need good actuators (Harmonic drives etc) for precision, and then good computing unit to handle all of it. Now you also need more power supply, and if it is mobile it needs a battery. Now your robot have a price tag only Arabic princes and well funded academic labs are interested in. And we haven't even touched the cost of try&fail iterative engineering process to make these work. In that process most companies/labs realizes the problem/condition have to be severely limited (run time, indoor/outdoor, general applicability vs made for one & one application only).

2. Human resource. Really, there's not much money in robotics, as a robotics engineer. Software robotics engineer can get a better working condition, security, salary & fulfillment in a software company. Mechanical, EE, embedded, etc is in a similar situation. Most robotics people I know are in for the passion. Application specific development (that is the norm right now) also require very niche knowledge that is hard to find.


The points you make are all true until they are not. Many useful robots should be able to work well enough for the purpose at a reasonable cost. Part of that is mass production brings prices down.

The problem is of course until we can solve the problem of making it work at all price isn't a consideration. You can go bankrupt advancing the state of the art which then is enough for someone else to take your work and (repeat the bankruptcy part many times until someone works it out), and finally someone makes a ton of money with their useful robot.


Servos are mass produced but still expensive!

You're right, but also software is weird. The world is absolutely littered with good computers that other people have already paid for, just waiting to be put to use for your project to expand their capabilities. If every new web app required its own little pocket computer to run, that you had to convince your customers to buy and keep in their home, it would be much less profitable to develop web apps.

The world is not littered with idle, high-performance robots waiting around for your motion planning algorithm to kick them into action. That makes it a lot more expensive and a lot less profitable.


Hmmm, lawyers, they will be involved. Add astronomical insurance and liability costs when things do go wrong in less-constrained and less-controlled environments.


> Two robotic hands that can work in co-ordination just as human hands do. Extremely difficult to solve, but money is there.

You will find the most automated Mc Donalds in the world in Switzerland. Mc Donalds over there has a lot more automation there than it does in the US, both in the cooking and in the order taking.

Either this is because the automation is extremely difficult to do, but is simpler in Switzerland. Or alternatively, this is because in most countries in the world humans are cheaper than the automated system, even after it has already been developed.

As someone who does research on robotics for a living. The problem is not the former, but the latter.


Can I surmise that most automation in dishwashing etc is geared for commercial enterprises ? I find the same lacking in a household. Sure you have the traditional dishwashers and vegetable choppers but they are largely clumsy to use and still take quite a lot of effort.

For a consumer, I don't know if the equation 'human labor cost << robots' holds true. It is a hassle to get human labor and there is no scope for time arbitrage. A robot can do the dishwashing job at night for example. Unlike a traditional dishwasher, you don't have to load the utensils. Just leave the utensils in a sink and you are done.


Two robotic hands that can work in co-ordination just as human hands do.

Rethink Robotics, a Rod Brooks startup, tried that, with "Baxter". Company went bust.

There's good, simple commercial hardware for high-volume vegetable chopping and dishwashing, and it can be found in most large commercial kitchens.

There's a huge amount of special purpose machinery in the world. Most of it is pretty dumb. Newer machinery tends to have more sensors and a bit of self-adjustment. Vision systems have become parts of much industrial machinery - they're cheap now. Mostly they're looking for things aligning or looking like a known good object.


They also had this obsession with low-bandwidth series elastic actuators, which make robots safer but slow and unproductive.

ABB's YuMi is perhaps a better example.


> Two robotic hands that can work in co-ordination just as human hands do. Extremely difficult to solve, but money is there.

Unless real human hands continue to cost less.


Human costs are complex. I used to work in a company that made tape robots. One customer did the math, hiring humans to change tapes was cheaper than buying a robot. Then he walked in one night and saw the kids in hockey gear, take the needed tape off the shelf, slap shot it to the "goalie" who put it in the drive. Which is why he was a customer even though human labor is cheaper.


as someone who has worked his share of minimum-wage jobs, this seems obvious. Perhaps it's a case of academics in their ivory towers not having sufficient exposure to the real world? That kid hired for $8/hr doesn't care about the success of the business. Their concern ends when the shift changes.


Robotics and embedded just doesn't make that much money. I had to turn down a job offer due to the seriously low counteroffer I was given. Or rather, I was given no counteroffer, just told to pound sand after the market rate I gave apparently insulted them.


I think a big part of the issue here is that the robotic boom hasn't happened yet. There just isn't that much demand for robotics skills, and the market is not efficient.

If robotics and machine learning can progress to the point where household robotics or general-purpose robots become a thing, there will suddenly be an explosion in demand. Salaries for robotics experts will go through the roof, and all of a sudden, everyone will want to study robotics in school. This could take another 10-20 years to happen though!


If you count autonomous driving vehicles as robotics (I know it's a stretch), the funding (and therefore pay) story is a lot better, although you get the usual startup vs BigTech debate.

Source: I work in an ADV company.


The amount of money that's gone into that area without shipping a product is insane.


The potential upside for the company that gets it right is enormous. Billions of people are tired of wasting their time driving. Entire industries can be built on the technology if it works well.

That said it's a problem that steers awfully close to needing a full real AI and that's been a showstopper for loads of potential solutions for decades now.


My observation is that the industry ("we") as a whole is slowly but steadily making progress, without throwing our hands and say we need AGI. A big part of this comes from the upscaling of test fleets, which now generate data with enough quantity and variety that you can leverage the existing data tools to analyze and give developers precise and actionable feedback. While Tesla might seem like a popular punch bag in the ADV world, they are a big practitioner of this paradigm and did get good value out of it. That, and the trend of replacing more and more hand-crafted rules with fuzzy numerical models designed to "blend together" the rules (therefore retaining explainability).


In some cases you need to invest as risk mitigation. If someone else does self driving cars and proves that it is safer than human driven cars by a large margin yiy might find yourself out of the car business when governments mandate the technology and you can't afford the license.


I would have thought the money wouldn't be there for a different reason - human hands suck compared to tools which is why we created them in the first place. It is a bit of a silly trope to have industrial work done by humanoid robots with hand tools.

The ability to manipulate destandardized sizes would be useful but precision manufacturing ate most of the lunch long ago which reduces it to more of a "last mile" task.


Well, the advantage I see with humanoid hands is that the same task that a human does can be done much faster by a humanoid hand. For instance chopping onions or other vegetables. Even dishwashing.


Kinda feels like that old life pro tip. Want to know the answer to something? Don’t ask a question, but make a false statement. Every expert in the field will rush to correct you and give you the most up to date and relevant information.


>Cunningham's Law states "the best way to get the right answer on the internet is not to ask a question; it's to post the wrong answer."[0]

Also I'd like to point out that you incorrectly stated it was an "old life pro tip" instead of an "internet law". Anyways, here is the correct information for you.

[0] https://meta.wikimedia.org/wiki/Cunningham%27s_Law


Thanks for setting me straight. :)


I don't really understand why depth estimation using binocular vision is still a problem.

I worked on this a bit a number of years ago and I thought I had scene matching working pretty well. The problem is I was trying to make it work without actually having two cameras (ie. on a smartphone where binocular cameras were not available at the time and for the most part still aren't). I was hoping to use the accelerometers and other sensors for dead reckoning the camera's position as the user swept it over a scene, but that turned out to be too hard.


Stereo vision isn't exactly "solved," but there are a ton of very good solutions out there both open source and commercial. The "Semi-Global Block Matching" algorithm implementation in the OpenCV library[1] is very good even though the algorithm is over 10 years old at this point. I've played around with the new Intel RealSense[2] units recently as well, which are stereo cameras with onboard processing and a nice IR pattern projector. Pretty cheap, and for the most part they "just work."

[1] https://docs.opencv.org/4.4.0/d2/d85/classcv_1_1StereoSGBM.h... [2] https://www.intelrealsense.com/stereo-depth


Isn't your "that turned out to be too hard" one of the answers?

People tend to think about machine vision in perfect conditions. This is the problem, me thinks. In practice, you will get a lens flare in most important case, or lose a sensor or get some rain.

Instead of designing solutions starting from most hardcore edgecases (which are common with humans), robotics researchers tend to provide MVP that works in best situation. That is far from end-user environmnent.

This needs global fixes, if you (I don't care) want robots working in real world.


> Isn't your "that turned out to be too hard" one of the answers?

That's a "we need an expert in filter theory" problem. I had four people crash and burn on that problem when we were building a DARPA Grand Challenge vehicle. Combining GPS, accelerometer, gyro, compass, and odometer data to get position is a hard problem. All those sensors are noisy, but in quite different ways. There are off the shelf solutions now, but there were not in 2004. We could not get below 3 degrees of heading noise, and had trouble keeping the sensor map aligned to the real world.


But what I said was that the "turned out to be too hard" part had nothing to do with binocular vision.

I'm certainly not claiming that I had a fully robust solution but I'm pretty sure lens flares wouldn't be too hard to filter out and there are lots of useful things that need doing in rain-free environments.


Do you think depth estimation could be done with two cameras plus computer vision (to find markers)? I think this is more or less what we do with our own eyes. Of course you would need much more processing power, but maybe for some applications the robot's brain doesn't need to be inside its body.


Parallax methods are widely used, I think it's how Tesla's driving assist features work.

The problem with it is that humans don't just use binocular vision - we have a whole model of the world. So for instance, if I see an object, I usually know roughly how big it is supposed to be because I have a conception of "object". I also know that the straight line on both sides of an object is a wall, and that the wall continues behind the object and therefore the object is in front of the wall and the object is closer than the wall.

That means I'm not just working with my binocular vision and it can be deceptive to think that because it works for human vision it will also work well for computer vision.


Lost binocular vision for a number of years back.

You get around just fine without it. Got it back after some speciality glasses (PRISM).

Was a complete shock see depth again. Didn’t seem to help anything getting it back. Mostly just trippy. Chairs were amazing to stare at.


From a much shorter and very different route, once after an 18 hour straight Quake marathon, I looked around the my room and was startled at how everything looked. I was highly impressed by the graphics and depth perception.


This reminds me of the 'Tetris effect' in which people who play Tetris for prolonged periods of time will begin to experience Tetris-like hallucinations when they stop. I've experienced it myself and have trouble explaining it, but it's as though everything you see becomes Tetris-like in some way. You look at your dinner and see ways to rearrange the peas so they 'fit' with the mashed potatoes, or something like that. There's more to it than that though, there's a real sensation of things being moving blocks that must be fitted together. It's bizarre.

But unlike Quake, Tetris isn't providing you with a 3D experience. That makes me wonder how much overlap there is between the two experiences.


I took an art class a few years ago. Negative space drawing and perspective drawing will blow your mind after a 3 hour session.

Walking outside you see lines and shadows everywhere. Entire perspective changes.


I’m curious to hear more. What caused you to lose and then regain depth perception? Were you able to see with both eyes or did you lose vision in one eye temporarily?


Mix of minor strokes and high intercraniel pressure.

My vision would switch from one eye to the other roughly every 30 seconds. Mostly seamlessly. Took ages to figure this out. Had some minor left brain issues. So writing would go gibberish every 30 seconds. Keeping a patch on right eye meant my vision went black every 30 seconds but my writing was fine when I could see.

Left eye being patched meant writing mostly gibberish. Still had blackness every 30 seconds.

Going on blood thinners resolved the vision going black and gibberish problems.

Prism glasses fixed double vision problems.

Medication for cranial pressure removes the need for prism glasses for about 10 hours. I have several extremely different pairs of glasses depending on what my brain is currently doing.

Inhaled some caustic gas a few years ago. My blood went “sticky” from it. Apparently it activated a Latent blood disorder.


This is the most interesting thing I've read today! Thank you for sharing!


Thanks! Took 3 years of hacking mind and body while mentally deficient to figure it all out. Leaned a lot from it all. Kind of forgot a lot too, but got most of it back now.

Would take memory tests daily or more to see how “stupid” I was.

Graphing my own cognitive decline and eventual resurgence was “interesting”


Whoa, I’m really sorry to hear that. Your story is fascinating and reminiscent of Michael Gazzaniga and others’ work with split brain patients.


If prism glasses helped regain depth perception, then probably something went wrong with eye motor function, causing diplopia (double image due to misalignment of eye). Could be trauma to the eye socket? Vision loss in one eye i think can be ruled out.


Sixth cranial nerve issue. I can keep eyes aligned, but is extremely difficult and painful if cranial pressure is high.

Reducing pressure resolves the issue.

Both eyes are very healthy.

At worse it’s +24 diapors, -1 power.

When pressure is good it’s 0-2 diapors +3power.

For awhile it was -6 diapors.

At this point I have glasses for much of the range.

Temple pain tells me when it’s time to switch glasses.

Avoiding caffeine and other stimulants helps.

Taking a Diamox will drop pressure in a couple of hours.


There is either a book or a great Joe Rogan podcast in your story. Thank you very much for sharing.


We can also go up and down abstraction levels at will.

We can see a forest, then a tree, then a leaf. All different "objects", yet related. The ability to not distinguish each blade of grass in my yard and just see it all as "grass" is very useful for noticing and not stepping on what the dog left behind.


Note that the human system fails for things like the moon which look vastly different in size at different times. This is proof of your point.


The answer is 'sometimes' and it's known as stereo vision.

Downsides include being less precise at longer distances (an object 1.5 meters away becomes a lot larger if it gets 1 meter closer, an object 150 meters away barely changes) and poor performance on surfaces with fewer features, or really dense features that all look alike (e.g. running across a field while spotting bumps and dips in the grass, or measuring whether sheets of steel are flat or not)

In some cases this doesn't matter - a Roomba doesn't care if it's hard to see things 150 meters away, as rooms are rarely that big.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: