The follow cam proof of concept is everywhere & nowhere. The problem is flying was only possible when someone else paid for it. Without any money, it's going to be a short effort.
Being a spare part application, 1 of the tiny servos from the Blade CX2 will be glued to the NTSC camera for pitch control. The Blade CX2 was an intriguing toy which never flew because it required a lot of trimming & its brushed motors were terrible.
The servo cam will be attached to the underside of a microquad. The microquad needs a better means of vibration damping, to use beat up propellers. It's getting near adding ballast like everyone else.
The camera will track a dual marker baton. The entire baton may actually be colored, making for a dumbbell marker that is easily isolated.
2 voltage regulators for powering the 5.8Ghz transmitter, & servo will be attached to the microquad. The camera is powered by the 5V flight computer rail, which originally was to power the PX4flow. The camera broadcasts video to the ground station, which combines it with telemetry to get a position for broadcasting back to the flight computer over bluetooth.
The ground station needs to record the video as well as scan it for position. The only spare part is a laptop. Since the follow cam's mane mission would be capturing running, the need to carry a laptop has been the showstopper.
For the 5 guys with jobs, the long reigning Odroid is finally dead. You need to be buying the Minnow Max. Unlike the boat, the Minnow has a dual core 1.3Ghz. It should be far ahead of the quad core ARMs. Personally have gotten nothing but grief from quad core ARMs.
Quad copter hoverbikes, autonomous follow cams, drone pizza delivery, self driving cars, self balancing motorcycles, & $5 trips to space are just around the corner, now. They've been just around the corner for a while, but now they're really just around the corner. Another impossible claim seems to come out every day, making all the stuff from last week obsolete.
The promises are getting bigger & bigger, but they're all extrapolating from the same old technology which hasn't improved in years. We've all been in a company that's trying to get bought out or have an IPO. They report every smallest, most tentative sale for the next 10 years, every possible source of revenue they could possibly make for the next 100 years, all as their current quarter's income.
That's what the current promises feel like, but every new promise extrapolates more than the last one. Merely breathing now nets $1 million in kickstarter donations & a belief that unicorns are going to ship in 3 months, just because unicorns breathe.
The current movement is feeding on the positive psychology movement that began in 2005. What began as the positive psychology movement mutated into refinancing ever less amounts of property into ever more amounts of money & that now has mutated into a requirement that every invention be believed, no matter how preposterous.
No-one asks questions anymore. No-one is allowed to have doubts. Pictures have nothing to do with what's...Continue Reading
With so many unemployed programmers, it was only a matter of time before someone would try to free the bits & upload some videos. We'll see how long this lasts until the indy movie studios crush the rebellion. These videos covered what currently are the 3 mane copters, in a rare appearance without any editing.
Collinn was a lot more down to Earth than when we was at DJI. It was still definitely a choreographed team demo, looking a lot like a Vegas trade show.
As with personal experience, the 1st attempt at a programmed flight was a hair raising flyaway & it took 2 attempts for it to work. He assumed it was a conversion from centimeters to meters, but the team never said what it was. When it did work, the droidplanner eye candy obscured what was happening, but you could barely see that it wasn't able to fly in a circle, in the wind.
They may have removed the circle from the flight plan because it wouldn't work. The user interface never showed it.
The eye candy made Droidplanner worthless outdoors, but it did show the potential for easy mission planning. It just needs a way to play back a simulation before flying, in order to avoid take 2.
20 years after it began, someone finally figured out how to route the audio output in ALSA back to an audio input. The sequence for recording from a Firefox player is:
Make Cinelerra record Loopback #1
The problem is the recording samplerate has to be the same as the playback samplerate. It has to be determined by trial & error or by playing the source in another player besides Firefox. If the player doesn't have a way to send audio to another device besides hw:0, you're screwed.
There is theoretically a way to do it with a line like pcm.!hw in the /etc/asound.conf file, but this never worked. pcm.!hw is supposed to replace any device called hw with what's described in the file.
Some time in 2012, someone wrote the snd_aloop module. The last ALSA release which could be easily compiled without upgrading the entire kernel was 1.0.25 & it has the snd_aloop module.
It creates a virtual soundcard for playback which really routes the audio to the recording handle of another virtual soundcard. aplay -l lists the virtual soundcard as
which translates to hw:1 for Firefox. They never quite figured out a naming convention, back in their dorm rooms in 1997, so there's a hodge podge of hw:1, Loopback #0, hw:1,0 naming conventions. Playing audio to Loopback #0 makes...Continue Reading
Searching far & wide for a marker system which might enable a follow cam revealed this:
Previous experience showed the classic international pink was the best chroma keying material & a doughnut was the easiest shape for current cameras to isolate. When chroma keying is combined with a unique shape, it gives 2 layers of redundancy.
A 4k camera detecting a full body, international pink, digital camouflage suit would be ideal.
Next would be digital camouflage markers, but given the reality of only having enough money for 640x480, we have only
Despite many flips & spinouts, it still managed to do it in 1h59m. Fastest mile was 8 minutes. Stopped the clock only for the battery change, but counted all the red lights & restarts. The battery died at 12 miles. Its speed rapidly dropped to 9m33s/mile. The new battery was back to 8m50s/mile. Battery life was 1h50m, giving a current consumption of 490mA with the headlights on. The right headlight constantly got bent down.
The human was powered by a salad & some sugar sticks. It made quite a bowel movement, but no crash.
So the bottleneck in making a viable follow cam is a super high resolution camera transferring realtime video to a really small computer where it's scanned for small, finely detailed markers. Interfacing HDMI has become the requirement for getting any kind of realtime video. You can probably get it into the raspberry pi's CSI bus. The CSI bus is a parallel bus of differential pairs. The pi isn't fast enough to do anything with it.
The only way to do the job is an FPGA implementation of the marker tracker, doing the full SDRAM, HDMI injestion, & marker detection in hardware. For all the hype about software, OpenCV, & running Pixhawk on Linux, hardware is still the only way to get anywhere near realizing the promises the UAV industry is making.
It's surprising no-one is focusing on hardware implementations of higher end object detection. There is a slow increase in hobbyist attention to FPGA's, but only for controlling LEDs, software radio, or very limited chroma keying from a low end VGA cam.
For all the different platforms they're trying to get Pixhawk to run on, they might as well port the autopilot to hardware. Everyone wants to be bought by Google & Google is a software company, so there's not a lot of thinking outside the box.
Another rover test ended when a motor fell off. Everything not threadlocked is being revealed when it falls off. Nothing was originally threadlocked because it wasn't supposed to work. It did 8m40s per mile when it worked. It was extremely hard to keep up. It flipped over twice on pavement cracks.
Accessing the offending bolt revealed all the sensors for the 1st time. The alignment is so critical that it didn't work when reassembled. Another sensor for the left motor had to be shifted ever so slightly closer. Fortunately, the error was a visible amount.
At 8.8V, with no headlights, the right motor now used 60mA more when going forward than when going backward, without load. Left motor used 20mA more when going forward.
At 15V, with no headlights, the no load current of a single motor was previously 250mA. The left motor was now 230mA going forward & 200mA going backward. The right motor was now 430mA going forward & 220mA going backward.
After reseating a bearing & another sensor reconfiguration, the right motor used 250mA going forward & 260mA going backward. This should have a significant impact on range. Another test run revealed it may now be too fast. So any direction problem is decidedly the mechanics rather than the sensors.
Key to the follow cam concept is object tracking & distance measurement. Just doing that without flying anything would be huge.
It's a little easier than full motion capture, since it only needs distance & position. It's a little harder, since there's only 1 camera. Doing that in daylight is still a hard problem. The last work Vicon showed was in 2011. It wasn't very convincing.
Lately, the Goog found a photo of a guy in a full body suit, covered in pattern markers. So obviously IR keying didn't work well enough. They replaced the reflective balls or LED emitters with more unique patterns a blob tracker could find.
Thresholding that didn't look good. The head & body might resolve, but the arm is gone. A really good piece of software could combine multiple cameras to get all the blobs, but a UAV only has 1 camera.
They also tried LED emitters on the previous sequel. What's needed is a better marker & high enough resolution camera so the marker can have enough unique features. Some bands of different types of hieroglyphics on multiple body parts might work. If 1 band is covered, it knows which body part another band belongs to, but not the distance between the bands. It may be that a tether is good enough to measure distance.
It's a fascinating problem on the cusp of being solved, but best left to the pros. There are so many people now pursuing UAVs, so many experts in machine vision, navigation, & Kalman filters compared to 3 years ago, it's solvable in short order.
You see it every time you boot up Linux. Along with Firefox & Open office, it's 1 of the 3 programs that have always defined the Linux desktop. It's The Gimp. It was truly extraordinary when it 1st came out. The user interface was better designed, tighter & more polished than any other free program at the time, by a wide margin.
It still crashed a lot, but it was your best hope of avoiding paying $300 for Photoshop. It couldn't swap images to disk for many years, making it require extremely large amounts of RAM, for the time. Other image editing programs like XV & Imagemagik were really hard to use.
The GUI toolkit it was written on became the standard for all the major desktop programs since, from Firefox to Google Chrome. It was ported to every platform. There was an extreme frenzy of development from 1996-1999 when Gimp, GTK, & most of the modern software world were being brought into existence.
The original authors of Gimp receded from the project but continued to be coworkers ever since & never had a rainy day, though their later work was never as famous as Gimp. They're now staff members at Square, which is considered the next Paypal & perhaps the next SpaceX.
Square was founded by the founder of Twitter. 1st encountered a Square terminal in 2012. It seemed like an obvious necessity that many banks were offering, but apparently Square was the only one. It was started in 2009, the optimum time to start a monopoly...Continue Reading
A better metric never discussed on the internets is how much a rocket can launch in a given timespan, rather than the amount it can launch in a single flight or the cost per mass in a single flight. The ground crew has to be paid constantly, no matter how many launches they send off.
Today's launch rates are glacially slow, compared to just 5 years ago. For all the difficulty in getting Falcon 9 to launch 3 times per year or Antares to launch once per year, it's hard to believe they were able to launch the monstrous shuttle 5 times in 2009.
The amount of mass launched in a fixed timespan is much lower today than it was 30 years ago. 30 years ago, they were putting 240,000 lbs of payload in low Earth orbit per year. It wasn't very good for the crew, but it was the highest capacity ever achieved. At $4 billion/year, it was $17,000/lb.
Today, Falcon 9 can put around 22,000 lbs per year in low Earth orbit. Antares can deliver 4400lbs per year. There aren't any figures for how much it's costing. If it was equivalent to the shuttle, Falcon 9 would be burning roughly $374 million/year. Antares would be burning roughly $74 million/year.
It actually doesn't sound far off what the rumors have been. There isn't any of the disclosure under the commercial programs that there was under the shuttle program. $57 million per flight could be a significant loss.
The 1st run with the brushless direct drive rover went way beyond predicted range, yet again. Range on 2S 900mAh was over 8 miles. Maximum downhill speed was 8 min/mile. Uphill speed sagged to 8m50s/mile. It was hopeless on rough terrain. It needed D feedback & shorter battery cables.
It was finally documented with the latest improvements.
The most expensive apartment in town is equivalent to the current situation, but 10 years behind in price. It has only 1 wall with windows, has open space, has a really solid front door. The most reasonable cheap place is another $300 cheaper, has 2 sides with windows, no open space. It's in a dump, has bars on the door with a separate lock, no bars on the windows, & has a really dumpy pool visible from the street. The difference is $3660/year. Between those 2 are other places which are extremely dumpy or not reasonable.
There's definitely a correlation between price & standard of living. It's not easy to endure a drop in standard of living & it seems worth spending zillions of dollars on happiness, which brings up the blond hottie paradox. Do unattainable middle age blond hotties spend zillions of dollars on expensive stuff because they can or because they're miserable? The other factor is if the price was sigificant, wouldn't you already have downgraded? Another choice of standard of living over price happened in 1998. It still seems impossible to say it wasn't worth it.
The mortgage & student loan booms have turned the west half of Bakersfield into an extension of LA, while the east half descends into the abyss. New construction, valley girls, & shiny black Audis straight off the 405 abound. The boom is far from the freeways, so it's not obvious to the Las Vegas trade show travelers, but a new superhighway seems destined to connect the new area to the 5.
The weather is difficult, but everywhere is difficult compared to the bay area. Eliminating sources of stress seems to be a noble priority.
The 1st road test with motor sensors worked perfectly. It definitely needed breaks & slightly faster startup. The stalls were gone. 7.4V got it to the 7.5mph range, without getting hot. Based on the rolling distance, it probably was near the no load current of 0.3A. It was virtually silent. Radio range was only 10ft.
It easily rolled over rough terrain. The rolling distance without breaks was surprisingly long. Even with no power, it rolled into curbs hard enough to flip over, but didn't flip over on rough road. The sensors seemed to withstand it. The castor wheel assemblies crumpled, reducing the impact.
Startup was a hard coded stepper acceleration, transitioning to commutating mode, to keep the wheels from slipping. That failed miserably. It slipped when rotating to the 1st step. The sensors couldn't make it transition into commutating mode seamlessly. Tried ramping PWM without a stepper state & it still lost steering. Carefully setting it to wait for the right stepper phase would probably solve the transition problem, but not the 1st step.
The only practical way to get straight driving was to go straight into full power. It didn't slip as much as feared. It used less current. Stepper acceleration is a dead topic. It consumes a lot of current for purely visual appeal.
After implementing heading hold steering & all the manual controls, it worked perfectly. All the issues with bang bang steering were gone....Continue Reading
This alignment was too widely spaced. The alignment had to be spot on to get reliable commutation. They had to be right on the edge of the magnets. The misaligned sensors gave 111, 000, or just 2 patterns with very small ranges for the other patterns. The aligned sensors gave all 6 patterns.
An experiment with derating the commutation time revealed very little derating could be done before the commutation time was too long to get useful back EMF. The working speed of the motor is just too low to use back EMF of any kind, even though it's fast enough to start it. It only starts because the commutation time is lowpass filtered & speeds up fast enough to get it into the sustainable range. The only solution is a sensor.
The loss of steering during startup was caused by the motors not changing state at the same time. Having the phase & state changes synchronized solved that problem until they stalled. At 15V, they still occasionally stalled when hitting rough pavement. A sensor would solve that & decrease the current.
There are no teardowns of a sensored motor, but blurry, shaky videos appear to show home made attempts with 3 hall effect sensors outside the motor casing. 1 guy used them in a modified e-bike. E-bikes are as common in the bay area today as Priuses were, 10 years ago. Of course, no-one has a clue how they work. That requires going to Hong Kong.
There are no examples of 2 sensored motors being used. It's a very complicated ordeal that makes people give up after converting 1.
A single A1321 sensor is applied to the outside & powered by 3.3V.
Decided to timelapse everything until the grand finale. Ran down with the guerilla pod, which was useless for video. At least the traffic was better than driving down with the video tripod. Being many years since the last manually exposed fireworks video, forgot everything.
1/30 F/2.8 ISO 1600 was a decent exposure. Stopping down would have sharpened the dots, but added noise. Desperately needed a video of a past show to know where to point the camera, but Sprint was completely down. Had to reposition during the show. 15mm with 1.6x cropping was just wide enough to get everything.
Audio was at the lowest setting. Clapping produced a decent approximation of the maximum loudness. Should have used a simple pair of outboard electret condensers, at minimum. Anything would have had better stereo separation than the Canon's microphone. Amazingly, it still managed to differentiate the guy on the right.
Now, they're doing it at slightly closer ranges, with modern wide angle, stabilized cameras. It has always depended on a very accurate attitude estimation with GPS coordinate triangulation.
The athlete still looks like a tiny dot. The newest videos are a lot more edited for when the athlete goes out of frame. It has the feel of trying to get a lot more mileage out of the same old capability.
Getting a closeup is really hard. The easiest way is to have a very long lens camera with many levels of stabilization & chroma key detection of the athlete. The last of the Sony Handycams had excellent stabilization. So far, they're all marketing gopro cameras, so this method is not being demonstrated any time soon.
The HDR-PJ540 is the current optically stabilized one. It's quite large & expensive. Stabilization on that level is now a novelty feature, since no-one cares. It's a huge investment, just for the follow me mode.
Another way is to fly up close. The advantage is a much easier time initializing it, much smaller vehicle, getting the athlete started in frame, & ability to fly in confined spaces.
GPS is no good. The athlete has to be the navigation reference. Time of flight cameras & structured light cameras don't work in daylight. The movie camera alone can resolve position, within strict limits. It would take having the athlete wear multiple chroma key markers & having 2 markers in frame at all time. The markers either have to be a constant distance apart or they need a way to tell how far apart they are, maybe by some electromagnetic sensor which has been demonstrated. Single camera autopilots have been done before.
It still would need GPS at long range. Any method using chroma keying without a GPS aid is going to be prone to false positives & flying away.